This article needs to be updated.(February 2022) |
An autonomous aircraft is an aircraft which flies under the control of automatic systems and needs no intervention from a human pilot. Most autonomous aircraft are unmanned aerial vehicle or drones. However, autonomous control systems are reaching a point where several air taxis and associated regulatory regimes are being developed.
The earliest recorded use of an unmanned aerial vehicle for warfighting occurred in July 1849, [1] serving as a balloon carrier (the precursor to the aircraft carrier) [2] Significant development of radio-controlled drones started in the early 1900s, and originally focused on providing practice targets for training military personnel. The earliest attempt at a powered UAV was A. M. Low's "Aerial Target" in 1916. [3]
Autonomous features such as the autopilot and automated navigation were developed progressively through the twentieth century, although techniques such as terrain contour matching (TERCOM) were applied mainly to cruise missiles.
Some modern drones have a high degree of autonomy, although they are not fully capable and the regulatory environment prohibits their widespread use in civil aviation. However some limited trials have been undertaken.
As flight, navigation and communications systems have become more sophisticated, safely carrying passengers has emerged as a practical possibility. Autopilot systems are relieving the human pilot of progressively more duties, but the pilot currently remains necessary.
A number of air taxis are under development and larger autonomous transports are also being planned. The personal air vehicle is another class where from one to four passengers are not expected to be able to pilot the aircraft and autonomy is seen as necessary for widespread adoption.
The computing capability of aircraft flight and navigation systems followed the advances of computing technology, beginning with analog controls and evolving into microcontrollers, then system-on-a-chip (SOC) and single-board computers (SBC).
Position and movement sensors give information about the aircraft state. Exteroceptive sensors deal with external information like distance measurements, while exproprioceptive ones correlate internal and external states. [4]
Non-cooperative sensors are able to detect targets autonomously so they are used for separation assurance and collision avoidance. [5]
Degrees of freedom (DOF) refers to both the amount and quality of sensors on board: 6 DOF implies 3-axis gyroscopes and accelerometers (a typical inertial measurement unit – IMU), 9 DOF refers to an IMU plus a compass, 10 DOF adds a barometer and 11 DOF usually adds a GPS receiver. [6]
UAV actuators include digital electronic speed controllers (which control the RPM of the motors) linked to motors/engines and propellers, servomotors (for planes and helicopters mostly), weapons, payload actuators, LEDs and speakers.
UAV software called the flight stack or autopilot. The purpose of the flight stack is to obtain data from sensors, control motors to ensure UAV stability, and facilitate ground control and mission planning communication. [7]
UAVs are real-time systems that require rapid response to changing sensor data. As a result, UAVs rely on single-board computers for their computational needs. Examples of such single-board computers include Raspberry Pis, Beagleboards, etc. shielded with NavIO, PXFMini, etc. or designed from scratch such as NuttX, preemptive-RT Linux, Xenomai, Orocos-Robot Operating System or DDS-ROS 2.0.
Layer | Requirement | Operations | Example |
---|---|---|---|
Firmware | Time-critical | From machine code to processor execution, memory access | ArduCopter-v1, PX4 |
Middleware | Time-critical | Flight control, navigation, radio management | PX4, Cleanflight, ArduPilot |
Operating system | Computer-intensive | Optical flow, obstacle avoidance, SLAM, decision-making | ROS, Nuttx, Linux distributions, Microsoft IOT |
Civil-use open-source stacks include:
Due to the open-source nature of UAV software, they can be customized to fit specific applications. For example, researchers from the Technical University of Košice have replaced the default control algorithm of the PX4 autopilot. [8] This flexibility and collaborative effort has led to a large number of different open-source stacks, some of which are forked from others, such as CleanFlight, which is forked from BaseFlight and from which three other stacks are forked from.
UAVs employ open-loop, closed-loop or hybrid control architectures.
Most UAVs use a radio for remote control and exchange of video and other data. Early UAVs had only narrowband uplink. Downlinks came later. These bi-directional narrowband radio links carried command and control (C&C) and telemetry data about the status of aircraft systems to the remote operator. For very long range flights, military UAVs also use satellite receivers as part of satellite navigation systems. In cases when video transmission was required, the UAVs will implement a separate analog video radio link.
In most modern autonomous applications, video transmission is required. A broadband link is used to carry all types of data on a single radio link. These broadband links can leverage quality of service techniques to optimize the C&C traffic for low latency. Usually, these broadband links carry TCP/IP traffic that can be routed over the Internet.
Communications can be established with:
As mobile networks have increased in performance and reliability over the years, drones have begun to use mobile networks for communication. Mobile networks can be used for drone tracking, remote piloting, over the air updates, [14] and cloud computing. [15]
Modern networking standards have explicitly considered autonomous aircraft and therefore include optimizations. The 5G standard has mandated reduced user plane latency to 1ms while using ultra-reliable and low-latency communications. [16]
This section needs additional citations for verification .(May 2016) |
Basic autonomy comes from proprioceptive sensors. Advanced autonomy calls for situational awareness, knowledge about the environment surrounding the aircraft from exteroceptive sensors: sensor fusion integrates information from multiple sensors. [4]
One way to achieve autonomous control employs multiple control-loop layers, as in hierarchical control systems. As of 2016 the low-layer loops (i.e. for flight control) tick as fast as 32,000 times per second, while higher-level loops may cycle once per second. The principle is to decompose the aircraft's behavior into manageable "chunks", or states, with known transitions. Hierarchical control system types range from simple scripts to finite state machines, behavior trees and hierarchical task planners. The most common control mechanism used in these layers is the PID controller which can be used to achieve hover for a quadcopter by using data from the IMU to calculate precise inputs for the electronic speed controllers and motors.[ citation needed ]
Examples of mid-layer algorithms:
Evolved UAV hierarchical task planners use methods like state tree searches or genetic algorithms. [19]
UAV manufacturers often build in specific autonomous operations, such as:
Full autonomy is available for specific tasks, such as airborne refueling [20] or ground-based battery switching; but higher-level tasks call for greater computing, sensing and actuating capabilities. One approach to quantifying autonomous capabilities is based on OODA terminology, as suggested by a 2002 US Air Force Research Laboratory, and used in the table below: [21]
Level | Level descriptor | Observe | Orient | Decide | Act |
---|---|---|---|---|---|
Perception/Situational awareness | Analysis/Coordination | Decision making | Capability | ||
10 | Fully Autonomous | Cognizant of all within battlespace | Coordinates as necessary | Capable of total independence | Requires little guidance to do job |
9 | Battlespace Swarm Cognizance | Battlespace inference – Intent of self and others (allied and foes). Complex/Intense environment – on-board tracking | Strategic group goals assigned Enemy strategy inferred | Distributed tactical group planning Individual determination of tactical goal Individual task planning/execution Choose tactical targets | Group accomplishment of strategic goal with no supervisory assistance |
8 | Battlespace Cognizance | Proximity inference – Intent of self and others (allied and foes) Reduces dependence upon off-board data | Strategic group goals assigned Enemy tactics inferred ATR | Coordinated tactical group planning Individual task planning/execution Choose target of opportunity | Group accomplishment of strategic goal with minimal supervisory assistance (example: go SCUD hunting) |
7 | Battlespace Knowledge | Short track awareness – History and predictive battlespace Data in limited range, timeframe and numbers Limited inference supplemented by off-board data | Tactical group goals assigned Enemy trajectory estimated | Individual task planning/execution to meet goals | Group accomplishment of tactical goals with minimal supervisory assistance |
6 | Real Time Multi-Vehicle Cooperation | Ranged awareness – on-board sensing for long range, supplemented by off-board data | Tactical group goals assigned Enemy trajectory sensed/estimated | Coordinated trajectory planning and execution to meet goals – group optimization | Group accomplishment of tactical goals with minimal supervisory assistance Possible: close air space separation (+/-100yds) for AAR, formation in non-threat conditions |
5 | Real Time Multi-Vehicle Coordination | Sensed awareness – Local sensors to detect others, Fused with off-board data | Tactical group plan assigned RT Health Diagnosis Ability to compensate for most failures and flight conditions; Ability to predict onset of failures (e.g. Prognostic Health Mgmt) Group diagnosis and resource management | On-board trajectory replanning – optimizes for current and predictive conditions Collision avoidance | Self accomplishment of tactical plan as externally assigned Medium vehicle airspace separation (hundreds of yds) |
4 | Fault/Event Adaptative Vehicle | Deliberate awareness – allies communicate data | Tactical group plan assigned Assigned Rules of Engagement RT Health Diagnosis; Ability to compensate for most failures and flight conditions – inner loop changes reflected in outer loop performance | On-board trajectory replanning – event driven Self resource management Deconfliction | Self accomplishment of tactical plan as externally assigned Medium vehicle airspace separation (hundreds of yds) |
3 | Robust Response to Real Time Faults/Events | Health/status history & models | Tactical group plan assigned RT Health Diagnosis (What is the extent of the problems?) Ability to compensate for most failures and flight conditions (i.e. adaptative inner loop control) | Evaluate status vs required mission capabilities Abort/RTB is insufficient | Self accomplishment of tactical plan as externally assigned |
2 | Changeable mission | Health/status sensors | RT Health diagnosis (Do I have problems?) Off-board replan (as required) | Execute preprogrammed or uploaded plans in response to mission and health conditions | Self accomplishment of tactical plan as externally assigned |
1 | Execute Preplanned Mission | Preloaded mission data Flight Control and Navigation Sensing | Pre/Post flight BIT Report status | Preprogrammed mission and abort plans | Wide airspace separation requirements (miles) |
0 | Remotely Piloted Vehicle | Flight Control (attitude, rates) sensing Nose camera | Telemetered data Remote pilot commands | N/A | Control by remote pilot |
Medium levels of autonomy, such as reactive autonomy and high levels using cognitive autonomy, have already been achieved to some extent and are very active research fields.
Reactive autonomy, such as collective flight, real-time collision avoidance, wall following and corridor centring, relies on telecommunication and situational awareness provided by range sensors: optic flow, [22] lidars (light radars), radars, sonars.
Most range sensors analyze electromagnetic radiation, reflected off the environment and coming to the sensor. The cameras (for visual flow) act as simple receivers. Lidars, radars and sonars (with sound mechanical waves) emit and receive waves, measuring the round-trip transit time. UAV cameras do not require emitting power, reducing total consumption.
Radars and sonars are mostly used for military applications.
Reactive autonomy has in some forms already reached consumer markets: it may be widely available in less than a decade. [4]
SLAM combines odometry and external data to represent the world and the position of the UAV in it in three dimensions. High-altitude outdoor navigation does not require large vertical fields-of-view and can rely on GPS coordinates (which makes it simple mapping rather than SLAM). [23]
Two related research fields are photogrammetry and LIDAR, especially in low-altitude and indoor 3D environments.
Robot swarming refers to networks of agents able to dynamically reconfigure as elements leave or enter the network. They provide greater flexibility than multi-agent cooperation. Swarming may open the path to data fusion. Some bio-inspired flight swarms use steering behaviors and flocking.[ clarification needed ]
This section needs to be updated.(February 2022) |
In the military sector, American Predators and Reapers are made for counterterrorism operations and in war zones in which the enemy lacks sufficient firepower to shoot them down. They are not designed to withstand antiaircraft defenses or air-to-air combat. In September 2013, the chief of the US Air Combat Command stated that current UAVs were "useless in a contested environment" unless crewed aircraft were there to protect them. A 2012 Congressional Research Service (CRS) report speculated that in the future, UAVs may be able to perform tasks beyond intelligence, surveillance, reconnaissance and strikes; the CRS report listed air-to-air combat ("a more difficult future task") as possible future undertakings. The Department of Defense's Unmanned Systems Integrated Roadmap FY2013-2038 foresees a more important place for UAVs in combat. Issues include extended capabilities, human-UAV interaction, managing increased information flux, increased autonomy and developing UAV-specific munitions. DARPA's project of systems of systems, [30] or General Atomics work may augur future warfare scenarios, the latter disclosing Avenger swarms equipped with High Energy Liquid Laser Area Defense System (HELLADS). [31]
Cognitive radio [ clarification needed ] technology may have UAV applications. [32]
UAVs may exploit distributed neural networks. [4]
An autonomous robot is a robot that acts without recourse to human control. Historic examples include space probes. Modern examples include self-driving vacuums and cars.
An unmanned aerial vehicle (UAV), commonly known as a drone, is an aircraft without any human pilot, crew, or passengers on board. UAVs were originally developed through the twentieth century for military missions too "dull, dirty or dangerous" for humans, and by the twenty-first, they had become essential assets to most militaries. As control technologies improved and costs fell, their use expanded to many non-military applications. These include aerial photography, area coverage, precision agriculture, forest fire monitoring, river monitoring, environmental monitoring, policing and surveillance, infrastructure inspections, smuggling, product deliveries, entertainment, and drone racing.
A micro air vehicle (MAV), or micro aerial vehicle, is a class of man-portable miniature UAVs whose size enables them to be used in low-altitude, close-in support operations. Modern MAVs can be as small as 5 centimeters - compare Nano Air Vehicle. Development is driven by commercial, research, government, and military organizations; with insect-sized aircraft reportedly expected in the future. The small craft allow remote observation of hazardous environments or of areas inaccessible to ground vehicles. Hobbyists have designed MAVs for applications such as aerial robotics contests and aerial photography. MAVs can offer autonomous modes of flight.
Swarm robotics is an approach to the coordination of multiple robots as a system which consist of large numbers of mostly simple physical robots. ″In a robot swarm, the collective behavior of the robots results from local interactions between the robots and between the robots and the environment in which they act.″ It is supposed that a desired collective behavior emerges from the interactions between the robots and interactions of robots with the environment. This approach emerged on the field of artificial swarm intelligence, as well as the biological studies of insects, ants and other fields in nature, where swarm behaviour occurs.
Aerial survey is a method of collecting geomatics or other imagery by using airplanes, helicopters, UAVs, balloons or other aerial methods. Typical types of data collected include aerial photography, Lidar, remote sensing and also geophysical data (such as aeromagnetic surveys and gravity. It can also refer to the chart or map made by analysing a region from the air. Aerial survey should be distinguished from satellite imagery technologies because of its better resolution, quality and atmospheric conditions. Today, aerial survey is sometimes recognized as a synonym for aerophotogrammetry, part of photogrammetry where the camera is placed in the air. Measurements on aerial images are provided by photogrammetric technologies and methods.
AeroVironment, Inc. is an American defense contractor headquartered in Arlington, Virginia, that designs and manufactures unmanned aerial vehicles (UAVs). Paul B. MacCready Jr., a designer of human-powered aircraft, founded the company in 1971. The company is best known for its lightweight human-powered and solar-powered vehicles. The company is the US military's top supplier of small drones —notably the Raven, Switchblade, Wasp and Puma models.
The Lockheed DC-130 is a variant of the C-130 Hercules modified for drone control. It can carry four Ryan Firebee drones underneath its wings.
The usefulness of UAVs for aerial reconnaissance was demonstrated to the United States in the Vietnam War. At the same time, early steps were being taken to use them in active combat at sea and on land, but unmanned combat aerial vehicles would not come into their own until the 1980s.
The Boeing A160 Hummingbird is an unmanned aerial vehicle (UAV) helicopter. Its design incorporates many new technologies never before used in helicopters, allowing for greater endurance and altitude than any helicopter currently in operation.
First-person view (FPV), also known as remote-person view (RPV), or video piloting, is a method used to control a radio-controlled vehicle from the driver or pilot's viewpoint. Most commonly it is used to pilot a radio-controlled aircraft or other type of unmanned aerial vehicle (UAV) such as a military drone. The operator gets a first-person perspective from an onboard camera that feeds video to FPV goggles or a monitor. More sophisticated setups include a pan-and-tilt gimbaled camera controlled by a gyroscope sensor in the pilot's goggles and with dual onboard cameras, enabling a true stereoscopic view.
The Australian Research Centre for Aerospace Automation (ARCAA) was a research centre of the Queensland University of Technology. ARCAA conducted research into all aspects of aviation automation, with a particular research focus on autonomous technologies which support the more efficient and safer utilisation of airspace, and the development of autonomous aircraft and on-board sensor systems for a wide range of commercial applications.
An uncrewed vehicle or unmanned vehicle is a vehicle without a person on board. Uncrewed vehicles can either be under telerobotic control—remote controlled or remote guided vehicles—or they can be autonomously controlled—autonomous vehicles—which are capable of sensing their environment and navigating on their own.
Ghatak is an autonomous jet powered stealthy unmanned combat air vehicle (UCAV), being developed by Aeronautical Development Establishment (ADE) of the Defence Research and Development Organisation (DRDO) for the Indian Air Force. The design work on the UCAV is to be carried out by Aeronautical Development Agency (ADA). Autonomous Unmanned Research Aircraft (AURA) was a tentative name for the UCAV. Details of the project are classified.
Paparazzi is an open-source autopilot system oriented toward inexpensive autonomous aircraft. Low cost and availability enable hobbyist use in small remotely piloted aircraft. The project began in 2003, and is being further developed and used at École nationale de l'aviation civile (ENAC), a French civil aeronautics academy. Several vendors are currently producing Paparazzi autopilots and accessories.
ArduPilot is an open source, uncrewed vehicle Autopilot Software Suite, capable of controlling:
Unmanned aircraft system simulation focuses on training pilots to control an unmanned aircraft or its payload from a control station. Flight simulation involves a device that artificially re-creates aircraft flight and the environment in which it flies for pilot training, design, or other purposes. It includes replicating the equations that govern how aircraft fly, how they react to applications of flight controls, the effects of other aircraft systems, and how the aircraft reacts to external factors such as air density, turbulence, wind shear, cloud, precipitation, etc.
INDELA-I.N.SKY is a Belarusian rotary wing unmanned aerial vehicle, medium-range with the weight up to 140 kg. It is developed and produced by KB INDELA Ltd. Serial production has been set up since 2014.
Primoco UAV is a Czech company that creates unmanned aerial vehicles designed for both civilian and military applications. Its first model took flight in July 2015 and the UAV Model One 100 started full production in January 2016.
The Raytheon Coyote is a small, expendable, unmanned aircraft system built by the Raytheon Company, with the capability of operating in autonomous swarms. It is launched from a sonobuoy canister with the wings deploying in early flight phase.
The HAL Combat Air Teaming System (CATS) is an Indian unmanned and manned combat aircraft air teaming system being developed by Hindustan Aeronautics Limited (HAL). The system will consist of a manned fighter aircraft acting as "mothership" of the system and a set of swarming UAVs and UCAVs governed by the mothership aircraft. A twin-seated HAL Tejas is likely to be the mothership aircraft. Various other sub components of the system are currently under development and will be jointly produced by HAL, National Aerospace Laboratories (NAL), Defence Research and Development Organisation (DRDO) and Newspace Research & Technologies.
{{cite web}}
: CS1 maint: multiple names: authors list (link){{cite web}}
: CS1 maint: multiple names: authors list (link){{cite journal}}
: CS1 maint: multiple names: authors list (link){{cite web}}
: CS1 maint: multiple names: authors list (link)