Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Next Article in Journal
Putting Temperature into the Equation: Development and Validation of Algorithms to Distinguish Non-Wearing from Inactivity and Sleep in Wearable Sensors
Next Article in Special Issue
Complex Electromagnetic Issues Associated with the Use of Electric Vehicles in Urban Transportation
Previous Article in Journal
Interictal Spike and Loss of Hippocampal Theta Rhythm Recorded by Deep Brain Electrodes during Epileptogenesis
Previous Article in Special Issue
Experimental Validation of LiDAR Sensors Used in Vehicular Applications by Using a Mobile Platform for Distance and Speed Measurements
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Autonomous Quadcopter Landing on a Moving Target

by
Alvika Gautam
1,
Mandeep Singh
2,
Pedda Baliyarasimhuni Sujit
3 and
Srikanth Saripalli
1,*
1
Department of Mechanical Engineering, Texas A&M University, College Station, TX 77840, USA
2
Robotic Institute, Carnegie Mellon University, Pittsburgh, PA 15213, USA
3
Department of Electrical Engineering and Computer Science, IISER Bhopal, Bhauri, Bhopal 462066, India
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(3), 1116; https://doi.org/10.3390/s22031116
Submission received: 30 December 2021 / Revised: 27 January 2022 / Accepted: 27 January 2022 / Published: 1 February 2022
(This article belongs to the Special Issue Feature Papers in Vehicular Sensing)

Abstract

:
Autonomous landing on a moving target is challenging because of external disturbances and localization errors. In this paper, we present a vision-based guidance technique with a log polynomial closing velocity controller to achieve faster and more accurate landing as compared to that of the traditional vertical landing approaches. The vision system uses a combination of color segmentation and AprilTags to detect the landing pad. No prior information about the landing target is needed. The guidance is based on pure pursuit guidance law. The convergence of the closing velocity controller is shown, and we test the efficacy of the proposed approach through simulations and field experiments. The landing target during the field experiments was manually dragged with a maximum speed of 0.6 m/s. In the simulations, the maximum target speed of the ground vehicle was 3 m/s. We conducted a total of 27 field experiment runs for landing on a moving target and achieved a successful landing in 22 cases. The maximum error magnitude for successful landing was recorded to be 35 cm from the landing target center. For the failure cases, the maximum distance of vehicle landing position from target boundary was 60 cm.
Keywords:
UAV; landing; guidance; vision

1. Introduction

Quadrotors are ubiquitous in numerous applications such as surveillance [1], agriculture [2,3], and search and rescue operations [4] because of their agile maneuvering, ability to hover, and ease of handling. During an autonomous aerial mission, landing of a quadrotor is the most critical task as any uncontrolled deviations can result in a crash. The task becomes even more challenging if unmanned aerial vehicle (UAV) has to track a moving ground target with minimum information while executing the landing maneuver simultaneously. This combined with time-critical requirements of certain applications and limited flight time of quadrotors makes accurate and timely completion of autonomous landing a challenging problem.
Autonomous landing of UAVs broadly involves two steps: (a) detection and/or knowledge of the landing pad (also referred to as the target) and (b) generation of motion control commands for the UAV to accurately land on the target. For the first component, i.e., target detection, landing using only GPS based target information is prone to errors of anywhere between 1 to 5 m [5,6] because of large covariance in measurements for localization. GPS is also unreliable in closed spaces and other GPS-denied environments. The use of onboard cameras and computer vision techniques for object tracking and localization is a popular way to overcome such shortcomings to enable autonomous landing not only on stationary but mobile targets as well [7,8,9]. Solutions for detecting the target using other sensors like light detection and ranging (LIDAR) [10], ultrasonic sensors, and IR sensors [11] are also available in the literature. Although interesting results were achieved, these approaches are not necessarily applicable to real-time outdoor applications involving moving targets. Both ultrasonic and IR sensors are sensitive to surface and environmental changes, and although LIDAR provides a very accurate estimate of distances, it is bulky and expensive for quadrotor applications.
Computer vision techniques to detect the landing target using onboard cameras was widely studied in literature [12,13,14,15,16,17]. Additionally, for landing on moving targets, a number of works have explored coordinated landing with active communication between the UAV and the target [18,19,20]. In this work, we do not rely on any communication between the UAV and the target to coordinate a landing maneuver.
Image-based visual serving is a popular approach for vision-based precision landing [21,22,23]. However, it relies on the landing target being visible throughout the task since it relies solely on visual information. To deal with scenarios such as intermittent loss of landing target, model-based approaches [24,25,26] were proposed to predict the trajectory of the landing target. Alternative solutions include attachment of additional sensors on the moving target (e.g., GPS receivers [27]). Additionally, some of the previously mentioned solutions rely on external computation such as communication between the UAV and a ground station for state estimation [25] and a Vicon system to deal with cases of no target detection [22], which is a limitation toward fully autonomous operation. To know the target location for plotting purposes, we mount a GPS on the target to record the trajectory of the target. We do not rely on an external motion capture system or active communication of information from target such as target location, velocity etc., to carry out the landing maneuver.
For the second component of autonomous landing, that is, generation of motion control commands based on target information, we will focus on the discussion of vision-based control techniques.
A large body of literature focused on traditional control theory approaches for landing [28] such as feedback linearization [29], backstepping control [30], mixed H 2 H [31] and fuzzy logic controllers [32,33,34,35]. Although these methods were extensively analyzed for stability, most of the testing was limited to simulation results.
Application of missile guidance laws to autonomous landing drew significant attention in the literature [36]. This typically involves the use of successive loop closure [37], which separates guidance and control into two separate loops, as shown in Figure 1. Guidance, in general, refers to the determination of heading from the vehicle’s current location to a designated target. The guidance commands are generated by an outer loop based on the target information. An inner loop controls the UAV attitude according to the generated guidance commands. It is desirable to use this approach as it is more flexible and offers finer control on each module without interfering with the working of inner proportional integral derivative (PID) control loops of the UAV and making it more flexible to integrate with different methods of target sensing .
Popular guidance approaches for landing include pursuit guidance laws such as nonlinear pursuit [38] and pseudopursuit [39], proportional navigation [27], and t g o guidance [40]. Although, guidance-based approaches are simple to integrate with vision in the loop while offering a finer control on the UAV trajectory, most uidance-based landing techniques proposed in the literature were tested for fixed-wing UAVs [41]. On the other hand, vision-based guidance techniques for quadrotor landing were limited to vertical landing approach only [42]. In such approaches, a quadrotor tracks the target while moving along a linear path till a certain distance, and then descends vertically to land on the target. This can result in increased flight time during the landing process. In this paper, we propose to use a simultaneous “track and descend” approach instead of a vertical landing approach for time efficient landing on a moving target using pure pursuit guidance. One of the obvious challenges of simultaneous track and descend approach is a faster decrease in camera field of view, which can quickly result in loss of the landing target. This warrants the need for a finer control on the speed profile of the UAV such that the UAV approaches and descends fast enough (for a time-efficient landing), while ensuring that the target does not go out of the field of view. To this end, we integrate the pure pursuit (PP) guidance law with a closing velocity controller.
Next, we list the contributions of this work.

Contributions

This work builds upon the authors’ previous work [43,44]. The work in [43] is validated in outdoor experiments for a stationary target and robustness of the approach for target tracking in the presence of environmental disturbances is presented in [44]. In this work, we extend the approach to moving targets. The major contributions of this work are as follows:
  • A fully autonomous PP-based guidance implementation with vision in the loop for time-efficient landing on moving targets, which descends and tracks the target simultaneously as opposed to track and vertical land approaches. The guidance approach is integrated with a log-polynomial function based closing velocity controller.
  • We add AprilTags to our landing target design and use AprilTag detection in addition to color-based object segmentation for detecting the landing pad. The proposed landing pad consists of AprilTags of multiple sizes and a logic is developed that switches between the two detection methods (blob detection and AprilTag detection) to accurately detect the landing pad from varying altitudes. A Kalman filter is used for target state estimation.
  • This work leverages the analysis of the controller’s parameter characterization based on different speed profiles presented in [45] for automatic initialization of the controller’s parameters, based on the initial estimate of target’s velocity. Thus, the approach does not rely on any prior knowledge of the target motion or trajectory.Also, the approach does not rely on any active sensor data being communicated from target to the aerial drone system about its state (position and velocity).
  • We demonstrate the performance of the proposed guidance law through realistic simulations for different target speeds and trajectories.
  • We evaluate the robustness of the approach through extensive real-world experiments on off the shelf 3DR solo quadrotor platform. We perform a total of 27 experiment runs, with scenarios consisting of straight line as well as random target trajectory along with scenarios of target occlusion.
  • Based on the findings of experiment results, we derive a a lower bound for vertical velocity of the UAV using the proposed controller so as to consistently maintain the visibility of the target so as to extend the approach for higher speed landing targets.

2. Materials and Methods

2.1. Landing Pad Detection Using Vision

Accurate and fast detection of the landing pad is essential. There are several visual patterns that can be used for a landing pad with different computational complexity [28]. In this work, we assume that the landing target is of known color (red) and no distractor targets are present in the vicinity. As the computational unit on board the vehicle is small, we use color segmentation and AprilTags [46] as the mechanism to detect the landing pad. Both these techniques have low computational complexity, and hence, can detect the pad at a high frame rate. Additionally, this approach does not require a particularly high image resolution, unlike approaches that require small object detection and accurate shape detection. Figure 2a shows the landing pad in the form of a red colored rectangle, with 16 AprilTags of varying sizes covering a portion of this red target. The red portion of the landing pad is detected using blob detection algorithm. The obtained RGB image in every frame is converted to HSV color space due to its robustness to change in lighting conditions. Hue defines the dominant color of an area, saturation measures the colorfulness of an area in proportion to its brightness, and the value is related to the color luminance.
Additionally, the original RGB image is also converted to YCrCb color space. In this space, color is represented by luma, constructed as a weighted sum of RGB values, and two color difference values Cr and Cb that are formed by subtracting luma from RGB red and blue components. A thresholding operation to detect red colored pixels was performed individually on both HSV and YCrCb color space frames.
For color segmentation, HSV performs better than RGB color space under varying lighting conditions. However, due to the nonlinear transformation involved in obtaining the HSV components from RGB results in nonremovable singularities (for black, white, and gray colors), which can cause issues. On the other hand, YCrCb components are obtained by a linear transformation of the RGB channels that is simpler, but due to the linear transformation there is a higher correlation between the component channels that can lead to false detection.
Through experiments, we empirically found that a mask obtained by combining both the HSV and YCrCb color segmented masks gave the best results for target detection in outdoor environment with nonuniform lighting conditions. Therefore, a bitwise OR operation was performed on the two obtained red pixel masks, and morphological operations were applied to remove the noise while preserving the final mask boundaries. Erosion was performed using a 8 × 8 rectangular kernel, followed by dilation using a 3 × 3 rectangular kernel. The red blob is detected by extracting the largest contour in the scene, as shown in the inset figure of Figure 2a.
During the terminal landing phase, camera footprint (m 2 area visible to camera at a specific altitude) is small, and hence, only a part of the target is visible or the complete frame is filled with the blob. At this point, our vision algorithm switches from detecting blobs to detecting AprilTags. The altitude at which this switching happens was 1 m in our experiments. Figure 2b shows the AprilTag selected by the vision algorithm. Such a landing pad design ensures flexibility and detectability from high as well as very low altitudes. AprilTags were just used for detecting the target center below a threshold altitude (similar to red blob detection at higher altitudes). We did not use AprilTags for pose estimation as that can be computationally expensive. The decision-making logic to switch between blob and AprilTag detection and AprilTag selection logic is explained in detail in Section 4. Additionally, we employ a target estimator to deal with cases of partial/full target occlusion as discussed next for efficient tracking of the target.

Target State Estimation Using Kalman Filter

During the landing process, the target may be partially or fully occluded, or may move out of the field of view. Hence, we use Kalman Filter (KF) to estimate the target parameters. The detection algorithm determines the x and y centroid pixel coordinates of the target, which is given as the input to the KF. The target is modeled in discrete time as,
X t ( n ) = F t X t ( n 1 ) + w ( n ) ,
where, w ( n ) is zero mean Gaussian process noise and the subscript n denotes the current time and n 1 denotes the previous time instant. X t is the state vector comprising of target characteristics: centroid pixel coordinates ( x t , y t ), rate of change of pixel coordinate positions ( x ˙ t , y ˙ t ), dimensions of the detected target contour ( w i d t h ( w c ) , h e i g h t ( h c ) ). The approach does not rely on distinguishing between width and height of the contour and instead utilizes the product (width*height) of the detected contour to give an insight into the area of the blob, since sudden changes in area may imply a wrong measurement or a false positive. F t is the state-transition model that is applied to the previous state X t ( n 1 ) . We measure the position of the target in image frame using vision. Measurement at time n is given by,
Z t ( n ) = H t X t ( n ) + v ( n ) ,
where, H t is the measurement model and v ( n ) is zero mean Gaussian measurement noise in landing pad parameter measurements. Both process and measurement noise are modeled as white, zero mean Gaussian noise.
The final estimated parameters of the target from the predict and update steps of the KF are given as input to the guidance loop.
Although we do make an assumption about the landing target being of a known color (red), overall, our landing target detection algorithm is reasonably robust to small red blobs in the environment, as the detection algorithm looks for the biggest contour as the landing target. Additionally, any sudden change in contour size or target center is mitigated by the Kalman filter estimation.

2.2. Log Polynomial Velocity Controller for Autonomous Landing

The guidance objective is to land the vehicle smoothly on the target while persistently tracking it. In this work, we use PP for autonomous landing control. PP works on the principle of consistently aiming to align the follower vehicle’s (i.e., the UAV) velocity vector towards the target. The guidance command for simple pure pursuit is given as,
a x y = V u λ ˙ x y K a ( χ u λ x y ) ,
where, χ u , χ t , λ x y , and R x y represent the UAV course angle, target course angle, line-of-sight (LOS) angle, and the distance between target and the UAV (range), respectively. V u is the UAV velocity and K a is the gain.
The details of 2D PP guidance and its extension to 3D for application to a landing scenario are described in Appendix A.
For a safe and precise landing, closing velocity (velocity with which the tracking vehicle closes on to the target) should reduce asymptotically to zero. This condition is mathematically represented as,
lim R 0 R ˙ 0 .
PP given in Equation (3) works under the assumption of constant UAV speed and thus does not control the UAV speed to drive the UAV target’s closing velocity to zero at the time of landing. To achieve the closing velocity requirement while landing in a time-efficient manner, we propose a log of polynomial function to control the vehicle speed. We integrate the speed controller with PP. The log of polynomial function is given as,
f ( x ) = N l o g n ( 1 + a x + b x 2 + c x 3 ) ,
N = V u 0 / V t 0 ,
x = R x y / R x y 0
subject to,
1 + a + b + c = n ,
where N is the ratio of initial UAV to target velocity and x is defined as the ratio of the current range of the UAV to target to initial UAV target range. UAV velocity is varied as,
V u = f ( x ) V t
where V u and V t are UAV and target velocity respectively. The variation of UAV velocity using a combination of pure pursuit and log polynomial speed controller assures the convergence of closing velocity to zero as the distance between UAV and the target (x) approaches zero during landing. Please refer to Appendix B for a complete proof of the same.
Besides achieving the desired closing velocity conditions, log of polynomial function has other desirable properties of the UAV flight as,
  • It has a slower decay for most of the flight, which makes sure that velocity of the UAV is significantly higher than the target.
  • Faster decay towards the end that quickly drives the closing velocity to minimum as the UAV approaches the target.
In 3D, R ˙ x y , λ x y reflect the velocity with which the UAV approaches the target and its alignment with the target. R z ˙ reflects the velocity with which the UAV descends. To satisfy the terminal landing condition given by Equation (4), it is important to drive both R x y ˙ and R z ˙ to zero. As explained in Theorem A1, R x y ˙ is driven to zero by driving velocity of UAV close to velocity of target. Driving R z ˙ to zero at the time of landing implies,
lim z 0 V z 0 .
To achieve condition given in Equation (10), we use an alternative log polynomial function for altitude control and this velocity profile is given as,
V z = V z 0 log n 1 + a x + b x 2 ,
where x represents the ratio of current to initial altitude, and V z 0 represents the initial velocity in z direction.
Most of the guidance-based landing techniques presented in literature [27,38,39] meet only the terminal constraints of landing without offering any explicit control over the velocity profile of the throughout the landing process. Although time-based guidance approaches exist that provide flexibility in terms of velocity profile control [40], they are prone to replanning requirements in case of system delays.
Our approach is distance- instead of time-based and overcomes some of these challenges as the solution design offers a more flexible control. Figure 3 shows the effect on the variation of parameters ( a , b ) on the velocity profile. For the sake of simplicity, we show the variation for two control parameters of the log polynomial instead of three. By varying ( a , b ) , we can control the landing velocity profile and various parameters like nature of the profile (initially accelerating or decreasing), maximum velocity attained, and time taken to land. Additionally, the theoretical analysis provided in [45] can be leveraged to serve as an automated lookup table to initialize as well as modify design parameters in different missions as well as during the same mission. This is a significant advantage as it not only provides a control on the terminal landing process but also allows to simultaneously incorporate multiple flight constraints (rate of velocity decay, maximum UAV, time to land) throughout the mission.
The decoupled nature of vision and control makes this approach simple and flexible to integrate with any distance based sensing modality. Further, the proposed controller allows for development of independent velocity control subsystems (Equation (9) for individual control of longitudinal and lateral velocity profiles based on target behavior), as well as ease of modification for vertical velocity profile control (11); therefore, it provides a practical as well as sound solution for real-time landing applications.

3. Simulation Results

We evaluate the performance of the developed vision-based PP guidance with log polynomial velocity controller using simulations on Microsoft Airsim and MATLAB, followed by experimental demonstration.

3.1. Simulation Setup

To evaluate the proposed guidance-based controller, we use the Microsoft Airsim [47] simulator using Unreal Engine [48] as the base. The vehicle simulated within airsim has a camera model (orientation and direction can be specified) and sensor models such as accelerometer, gyroscope, barometer, magnetometer, and GPS. Onboard stream (of 640 × 480 resolution) of the downward camera was used for target detection. The landing target is in the form of a moving car simulated in unreal engine. We use the blob detection approach discussed in Section 2.1 to identify the landing target. Figure 4 shows a screenshot of the simulation environment consisting of the flying quadrotor vehicle modeled in Airsim and a ground vehicle, which served as the moving landing target.
We evaluated the proposed strategy through airsim simulations for a straight-line moving target for 3 different speed cases and a circular trajectory moving target for two different minimum turning radii. In each scenario, the quadrotor was initialized at an altitude of 20 m.

3.2. Landing on a Straight Line Moving Target

In this case, we evaluate the performance of the system when the target is moving in a straight line motion with a constant speed. Simulations with different target speeds of 1 m/s, 2 m/s and 3 m/s are carried out.
Figure 5a–c shows the trajectory followed by the quadrotor with the vision-based target input when the target is moving with a constant speed of 1 m/s, 2 m/s, and 3 m/s respectively. With vision in the loop, the quadrotor follows an efficient trajectory towards the target.
Figure 6a,b show the closing velocity profile and the descent velocity of the UAV respectively for 2 m/s target speed case. It can be observed that both converge to zero as the vehicle lands on the target.

3.3. Landing on a Circular Trajectory Moving Target

In this case, we analyze the performance of the quadrotor landing when the target is moving in a circular trajectory of radii 5 and 15 m with a speed of 2 m/s.
Figure 7a shows the trajectory followed by the vehicle to successfully land on the target for 15 m radius case. As observed from the figure, the quadrotor tracks the trajectory of the target vehicle and efficiently adjusts itself to follow along the same path and finally land on the target accurately.
Figure 7b,c show the closing velocity profile and the descent velocity of the UAV respectively. Both converge close to zero as the vehicle lands on the target. Figure 8 shows a similar trend of trajectory, followed by successful landing when the landing target was made to follow a circular path of a smaller radius (5 m approx.). Since the target maneuvers rapidly, the vehicle takes a longer duration to land in this case. For the 5 m radius case, the aerial vehicle descends slower than the straight line case since the longitudinal and lateral controllers have to keep the small radius target in the field of view.

3.4. Matlab Simulation Results: Landing with Noise in Target Information

In the previous case, it was assumed that accurate information about the target is available to the UAV. In the field, the received target information is often noisy. Therefore, to analyze the robustness of the landing controller simulations are performed with a noisy target model.
Noise was added to the target position in the form of measurement noise ( N ( 0 , 1 . 5 2 ) ) and to target velocity in the form of process noise ( N ( 0 , 0 . 4 2 ) ), both were considered to be non correlated and zero mean white Gaussian noise. Figure 9a,b show the estimated target trajectory and the simulated trajectory followed by the UAV to land on the target. UAV was able to land on the target in the presence of noise while satisfying the desired terminal conditions.

4. Experimental Results

We use an off-the-shelf quadcopter (3DR SOLO) for hardware validation of our landing algorithm. The landing algorithm runs on a Linux-based embedded computer–Odroid U3+. This is mounted at the bottom of the quadrotor and connected to a tilted Ptgrey chameleon USB camera. In addition to the camera, we also use a Sf02 Lidar connected to the Odroid via USB for better altitude estimates. These altitude estimates are used only by our controller running on odroid. Also, although a Lidar was mounted on the drone for better altitude estimates, this altitude estimate was not dependent on the presence of the target but rather the height of the drone above ground. Thus, though this altitude estimate helped in better estimate of target parameters this information by itself was not a result of active communication between the UAV and the target.
Output of our controller is in the form of velocity commands that are sent to the autopilot, where corresponding attitude commands are generated by the low level control loops in the autopilot. These attitude commands are finally converted to motor signals. Details of the hardware setup and the information flow are shown in Figure 10a. The proposed landing methodology is implemented using ROS [49], and OpenCV [50] is used to implement target detection. Figure 10b shows the quadrotor above the landing target. The inset Figures show 3DR solo with all the associated accessories and image obtained from the onboard camera, with the landing target detected by the vision algorithm. The communication between the onboard Linux computer (odroid) and the drone’s autopilot was established by connecting odroid to 3DR SOLO’s own WiFi network. Sensor data from the autopilot (drone’s compass heading etc.) is received by the onboard computer and velocity commands generated from the onboard computer are sent to the drone using this WiFi link. The onboard computer sends and receives information in the form of ROS messages, whereas the drone’s autopilot exchanges information in mavlink format. Thus, we use a standard ROS package mavros that acts as a bridge for message format conversion from ROS to mavlink and vice versa, thus establishing seamless exchange of information between the drone and the onboard computer. The videos of the landing experiments are available at Supplementary Materials.

4.1. Landing System Architecture

Landing system architecture consists of vision and control pipelines as shown in Figure 11. Vision processing pipeline is composed of detection and estimation submodules. Detection module consists of a combination of blob detection and AprilTag detection as explained in Section 2.1 and the estimation module consists of a Kalman Filter. Whenever the target is detected, appropriate velocity commands are published by the controller to the autopilot based on vision information. If no measurement is received by the estimator (target is not detected), guided velocity commands are published based on the target position predicted by the estimator, and a no-measurement time counter is started. If the target is not detected continuously for 6 s, V x = 0 , V y = 0 , and V z = 0 velocity commands are published for x , y , z resulting in a state of hover until the target is detected again. The value 6 s was selected after several experiments. This transition of mode prevents any crashes or undesired maneuvers in cases of target occlusion and/or target going out of the field of view.
Next, we discuss the Trigger Landing logic during the terminal landing phase. As mentioned in Section 2.1, the camera footprint decreases with altitude and becomes very small during the final landing stage. Therefore, it is challenging to keep the entire blob in the camera frame, and thus only a part of the landing target might be visible. In such a scenario, only blob detection may not be sufficient to accurately detect and land on the target center. To this end, we place a set of 16 AprilTags of varying size inside the red colored landing target and the vision logic switches from blob detection to AprilTag detection below 1 m altitude. Multiple AprilTags with varying sizes make it easier to detect at least one AprilTag to aid the Trigger Landing process, during the final stage. Additionally, AprilTags are translation, rotation, scale, shear, and lighting invariant, resulting in robust detection. Below 1 m altitude, if at least one AprilTag is detected, it is given preference over the detected blob and AprilTag center is output by the vision pipeline as the landing target center. In the case of multiple AprilTag detection, estimator tracks the AprilTag selected in the previous frame (using the tagID information), for a smooth motion. If none of the detected AprilTags IDs match, the ID of tag detected in the previous frame, the vision algorithm does a random selection of one of the tags as the target center. During the trigger landing process, if at least one AprilTag is detected in every frame continuously for 3 s, then the controller triggers the landing action giving zero velocities in lateral and longitudinal direction and high value of downward z velocity (to compensate for ground effect). After touchdown, the vehicle is taken into manual control and disarmed. This approach of allowing the vehicle to touchdown followed by disarming and motor shutdown was an experimental design decision, as shutting off the motors directly below a threshold altitude can sometimes result in toppling of the vehicle and/or harm to onboard equipment.

4.2. Landing on Moving Target

Next, we evaluate the log polynomial controller for a moving target scenario. We conducted a total of 27 experiment runs to test the efficacy of the proposed strategy. The considered experiment cases included: landing on a target moving in a straight line, random trajectory and landing on a moving target with temporary midflight full target occlusion. We conducted 9 trials each for first two cases, and for the last case of mid flight full target occlusion, 5 trials with straight-line moving target and 4 trials with target moving in random trajectory were conducted. Details of these trials are presented in Section 5.

4.2.1. Straight Line Target Motion

For this case, the quadcopter was commanded to takeoff to an altitude of 10 m and the target was moved in an approximate straight line motion with an approximate velocity of 0.3 m/s (as given by the autopilot mounted on the target) in the longitudinal direction and negligible movement in the lateral direction. In case of a moving target tracking and landing, the velocity of the UAV must be controlled according to the velocity of the target. Thus, it is important to compute the target velocity correctly. We use a KF for a robust estimate of the landing pad parameters ( x i m g , y i m g ) , to deal with sudden changes in detection output, resulting from wind disturbances and target occlusion. The relative velocity on the image frame is calculated by the slope of the image positions. We can express the target velocity as given below:
v ( t x ) v ( t y ) = v q x t δ ( t ) v q y t δ ( t ) + K x 0 0 K y cos ( ψ ) sin ( ψ ) sin ( ψ ) cos ( ψ ) x i m g ˙ y i m g ˙ .
where ( K x , K y ) , represent the constants for transformation between the local NED coordinate and the image plane. The computed target velocity components are used to design the lateral and longitudinal velocity control subsystems independently using the log polynomial velocity profile. Altitude control was achieved using the velocity profile given in Equation (11), since the requirement in z direction during the terminal landing phase is V z 0 , as z 0 . For x y plane velocity control, since the aim is to drive the UAV target’s relative velocity to zero velocity, the profile remains the same as mentioned in Equations (5)–(7). Further, a flag was set for the quadrotor to descend such that quadrotor was commanded a z velocity only if this flag was equal to 1 as given in Equations (13)–(15),
f l a g d e s c e n d = 1 , if flag detected = 1 ,
V z = V z 0 log n 1 + a x + b x 2 , if flag descend = 1 ,
V z = 0 , if flag descend = 0 .
The flag was set to 1, if the target was detected by the vision system. In case of no target detection, commands for tracking the target based on estimates were published. We conducted several trials to evaluate the efficacy of our controller. Details of these experiments are presented in Section 5. Next, we discuss the results from 2 trials (trial 1 and 2) for landing on a target moving in a straight line. Figure 12a shows the trajectory followed by the quadrotor to land on the target during trial 1. Figure 12b–d show the longitudinal, lateral and descend velocity profiles of the quadrotor respectively for this trial. The longitudinal velocity profile approaches the target speed toward the end and the lateral velocity profile approaches 0, thus resulting in zero closing velocity for a successful landing.
Figure 13 shows results of trial 2 straight-line moving target . For this trial, a different set of log polynomial parameters were generated randomly ( ( a , b , c ) = ( 55 , 115 , 161 ) ). Initial conditions for this experiment (UAV-target distance, initial UAV velocity, target velocity) were same as that of trial 1. Figure 13b shows that a faster landing is achieved in trial 2. This is because of a higher value of log polynomial parameter a in trial 2 ( a = 55 ) resulted in a higher initial acceleration compared to that of trial 1 ( a = 40 ) . We show only the longitudinal velocity profile for trial 2 as significant motion of target is in longitudinal direction only. Figure 14 shows the external camera view of a typical landing sequence. Inset figure in every frame shows the onboard camera view.

4.2.2. Target Moving in a Random Trajectory

In addition to a straight-line moving target, we also performed hardware experiments for target moving in a random trajectory with an approximate speed of 0.4 m/s. Figure 15a shows the UAV-target trajectory for such a scenario. Figure 15b–d show the lateral, longitudinal, and z velocity profiles respectively for the trajectory in Figure 15a. In both cases, UAV is able to land successfully on the target with UAV velocity eventually converging to target velocity before the landing is triggered. Figure 16 shows another scenario for a nonstraight-line moving target landing. The altitude jump in Figure 16 is attributed to the spike in LIDAR data due to a slight change in ground slope.

4.2.3. Landing on a Moving Target with Midflight Occlusion

One of the major challenges of the landing process is losing the sight of the target. This can happen if the target gets partially or fully occluded due to obstacles and/or goes out of field of view. It results in a deviation from the desired trajectory. To overcome this situation, we use the KF to achieve a robust estimate of the landing pad parameters. The output of detection algorithm ( x t and y t centroid pixel coordinates of the target) are given as input to the estimator. The extracted target parameters using vision are used to generate appropriate guidance commands for the quadrotor.
For this experiment case, we fully occluded the target during certain phases of the flight. Figure 17 shows the results for one of the instances where the target was manually occluded. In Figure 17a, a zero value of detected flag indicates that the target is occluded and/or out of field of view. When the target is occluded, no measurements are received, and the predicted estimate by KF is used for 6 s (empirically chosen) to generate the guidance commands. This is represented by the value of 1 for tracking flag even when the detection flag is 0. After 6 s, zero velocity commands are published by the controller and tracking flag is set to 0, until the landing pad is detected again. Figure 17b shows the trajectory followed by the UAV to land on the target in midocclusion flight. Since the target was recovered after the occlusion, the vehicle is able to land on the target successfully.
Figure 17c,d show the altitude and z velocity profile, respectively. Whenever the detection flag is zero, V z is also zero, thus maintaining a nearly constant altitude that results in a nondecreasing camera footprint, making the target reacquisition easier. The second instance of target leaving UAV camera view (just after the 40 s time mark) shows a sudden jump in altitude in Figure 17c. Although the corresponding commanded z velocity in Figure 17d is zero. The sudden spike in altitude was the result of an atypical behavior of the autopilot due to commanded z velocity being zero momentarily at very low altitude. For this specific case, we were able to reacquire the target because of the altitude spike. In the future, we plan to incorporate a hybrid approach for similar scenarios. As a part of the hybrid approach, we plan to include a search strategy where, if the target is lost for more than a prescribed threshold period, the controller will command an increase in UAV altitude to aid search, and hence, the target reacquisition.

5. Discussion

As discussed in Section 1, significant work was performed in literature for autonomous landing of a quadrotor on moving targets. In [27] authors proposed a guidance-based approach for autonomous landing. The approach consisted of PN (proportional-navigation) guidance during the approach phase augmented with a closing velocity controller and proportional-derivative (PD) controller for the final landing phase. However, the closing velocity controller was not tested with vision in the loop and was only limited to approach phase with UAV flying at a constant altitude. Such a closing velocity controller may not work for landing when there is a simultaneous decrease in altitude. Although experiments showed landings at higher target speeds, their approach relied on target information from inertial/global positioning system(GPS) sensors placed on the ground vehicle, in addition to vision. In [51], the tracking guidance for the vision-based landing system was based on relative distance between the UAV and the target. However, the decrease in altitude during the landing process was a step-wise decrease, i.e., a downward velocity was commanded intermittently only when the UAV was directly above the target and UAV was in target tracking mode (at constant altitude) at other times. Such a noncontinuous decrease contributed to an increase in total landing time. This is one of the examples that highlights the advantage of our approach, where a common log polynomial speed controller function (albeit with different parameter values) for horizontal and vertical speed control provides explicit control over the speed profile characteristics resulting in a time-efficient landing. From an evaluation perspective, most approaches in literature reported the results in the form of aggregate landing accuracy from a limited number of trial runs [7,27,52]. We improve on the existing work through a more extensive evaluation. We conduct a total of 27 outdoor field experiments consisting of straight line as well as random target trajectory cases (with and without intermittent target occlusion). Further, we report the average landing time of each trial category in addition to landing accuracy. Next, we present the details of the experiment runs along with the observations and lessons learned from failure landing cases.
The criteria to evaluate whether a trial was successful was the distance of the vehicle from the target center. A trial with less than | 40 | cm distance between the vehicle’s final landing position and target center was categorized as successful; otherwise, unsuccessful. Table 1 shows the categorization of experiment trials along with the details of landing accuracy and the number of unsuccessful trials. As per the criteria mentioned above, we were able to achieve a successful landing in 22 runs. In 13 of these runs, we were able to achieve a landing within ± 20 cm from the landing target center. For the rest of the 9 experiments, the landing was achieved within ± 35 cm of the target center. The table also shows the average landing time and best landing accuracy (least error between vehicle landing position and target center) achieved in each category. During the successful trials, the fastest landing was achieved in 41 s in the straight-line moving target scenario with the log polynomial parameter values of ( a , b , c ) = ( 61 , 109 , 161 ) with a landing accuracy of | 0.12 | m. Maximum and average UAV speed for this case in the direction of target motion was 0.67 m/s and 0.5 m/s, respectively. Log polynomial parameter values of a > 61 , although provided a higher initial acceleration of UAV velocity resulting in a faster landing, the landing accuracy suffered. This was in all likelihood due to a rapid deceleration and ungraceful decay of closing velocity during the final stages of the landing process. Figure 18a shows a boxplot for successful experiment trials. Y-axis represents the observed distance between target center and vehicle landing position during these trials and the 4 boxes correspond to the four experiment scenarios discussed in Table 1. The boxplot shows that the median of error for successful landing trials is well below 0.2 m for all the cases, except random trajectory with target occlusion where it is 0.30 m. Figure 18b depicts the vehicle landing position for different experiment runs. Green markers and yellow markers show the cases for vehicle landing within ± 20 and ± 35 cm from the landing target, respectively. White markers represent the failure cases where the vehicle failed to land on the target.
We had 5 unsuccessful runs where the vehicle landed away from the target. Among the unsuccessful trials, the maximum observed vehicle distance from the target was 0.6 m, which occurred in the “random trajectory moving target with mid-flight occlusion” case. During this trial, occlusion was introduced at approximately 4 m of altitude. Additionally, the trajectory of the moving target was perturbed, while the target was occluded. Thus estimator (in the absence of measurements) was unable to predict the target motion correctly during the occlusion phase.
It was observed during the trials that all the failures occurred due to target going out of field of view, because of (i) erroneous estimation of target motion at low altitudes during occlusion and (ii) perturbations in target trajectory, especially at the time when quadcopter was slightly lagging behind the target. These failures can be attributed to an increase in the relative estimation error of the KF at lower altitudes (because of a decrease in camera footprint). Instances of the vehicle lagging behind the target due to estimation errors were also observed during successful trials. However, the controller was robust and quick enough to catch up as soon as the target was detected, and the state was updated correctly.
For all the experiments, a strictly decelerating profile was chosen for z-direction since a smooth decrease in camera footprint is desirable for maintaining target visibility. Log polynomial parameters for lateral and longitudinal velocity profiles were initialized based on the estimated target motion. For example, for the straight-line moving target case, an initially accelerating profile followed by deceleration was chosen for longitudinal direction, since the significant target motion was limited only to this direction only. In the case of random target trajectory landing, an initially accelerating profile was chosen for lateral as well as longitudinal direction since the target had significant motion in both the directions.
From our experiment trials, we inferred that any environment disturbances such as wind, or any abrupt changes in target trajectory results in target going out of the camera view. To this end, we evaluated conditions for descending velocity to maintain target visibility more consistently.
Consider an aerial scenario as shown in Figure 19, where a quadrotor (with a velocity V u ), and with a downward facing camera, is attempting to land on a moving target with velocity V t . At time t, quadrotor is at a distance R from the landing target and an altitude of h meters above the ground.
Let D meters be the distance (in meters) present in the camera field of view at a height h above the ground at time t. The condition of keeping the target in field of view can be expressed as, R D , R ˙ < 0 . Given such a scenario, the target remains within the field of view of the UAV, as long as the velocity of UAV is higher and/or consistent with the target velocity and a near zero relative velocity is maintained in the final stages of landing. With known intrinsic camera parameters and quadrotor velocity profile, a bound in descend velocity can be calculated so as to maintain target visibility.
Given camera parameters (field of view in degrees), and log polynomial controller given by Equation (5), the condition for maintaining target visibility can be expressed as:
V z < 2 V t N 1 tan ( f o v 2 ) ,
where f o v is field of view of the camera in degrees, N is the ratio of initial UAV to target speed, V z and V t is descend velocity and target speed, respectively. The current experiment trials were performed at low target speeds. While we would have liked to test our controller at higher speeds, a combination of factors like our system configuration, along with the added payload and safety restrictions of the available outdoor experiment space, constrained us to limit our experiments to lower speed moving targets.
In addition to driving the closing velocity to zero and achieving a good landing accuracy, one of the significant challenges for landing on a fast-moving target is to maintain its visibility consistently. Based on the lessons learned from our experimental trials, for landing on a higher speed moving target using our proposed framework, we anticipate more challenges for circular and random target trajectories (with and without occlusion) than a straight-line target trajectory.
For the scenarios mentioned above, a robust estimation of target motion is generally one of the major concerns. For example, in our case, the camera frame rate of 15 Hz is most likely too less for a target moving in a random trajectory with a speed of 10 m/s. Since our estimate relies on continuous target measurements to correct its estimate, prediction uncertainties would be higher with a low update rate of target measurements and much higher during target occlusion. Although in our previous work we performed an extensive analysis of different velocity profiles with variation in log polynomial parameters, a choice of these parameters for landing on a high-speed moving target in the presence of estimation errors would be another challenge, as any sudden change at high aerial vehicle speed is likely to cause aggressive motion of the drone. Thus, we intend to develop an automated method for log polynomial parameter selection during the landing process and use the result presented in Equation (16) to further improve the robustness of our approach for landing on higher speed moving targets.

6. Conclusions

In this paper, a guidance-based landing approach is presented for autonomous quadrotor landing with vision in the loop. A log-polynomial closing velocity controller is developed and integrated with pure pursuit (PP) guidance for a time-efficient landing with minimum closing velocity. The integrated approach enables the vehicle to land faster compared to that of the traditional track and vertical landing. The efficacy of the proposed approach is validated exhaustively through simulations. We also perform extensive outdoor field experiments to validate the proposed approach for a stationary and moving landing pad using a quadrotor with onboard vision. The results show that the quadrotor is able to land on the target under different environmental conditions. The landing experiments were carried in mild wind conditions. To explore the ability of the quadrotor to land under high wind conditions, one may have to design observers and nonlinear controller that are robust to these disturbances. A potential future work involves a more sophisticated mechanism of choice of controller parameters at the beginning and midflight to deal with high speed, as well sudden accelerations/change in target motion. Another direction of work could be to develop novel landing controllers for ship-based landing where the target has a 3 D motion and the environment has wind disturbances, which are a function of the sea state. One could also explore different sensor fusion approaches for landing and show a comparison of different approaches for landing.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/s22031116/s1.

Author Contributions

Conceptualization, A.G., P.B.S., S.S.; methodology, A.G., P.B.S., S.S.; software, A.G.; validation, A.G., M.S.; formal analysis, A.G.; investigation, A.G., M.S., P.B.S., S.S.; resources, P.B.S., S.S.; data curation, A.G., M.S.; writing—original draft preparation, A.G., M.S., P.B.S., S.S.; writing—review and editing, A.G., P.B.S., S.S.; visualization, A.G.; supervision, P.B.S., S.S.; project administration, P.B.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was partially funded by EPSRC EP/P02839X/1.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Basics of Pure Pursuit (PP) Guidance

In this section, we provide an overview of 2D and 3D pure pursuit guidance laws for completeness, and then use these guidance laws for landing in the subsequent sections.

Appendix A.1. 2D Pure Pursuit (PP) Guidance

Consider a planar UAV-target engagement geometry as shown in Figure A1. The speed of the UAV and the target is V u and V t , respectively. The notation χ u , χ t , λ x y , and R x y represent the UAV course angle, target course angle, line-of-sight (LOS) angle, and the distance between target and the UAV (range), respectively. The target is assumed to be nonmaneuvering; that is, χ t is constant. Using the engagement geometry in Figure A1, R ˙ x y and λ ˙ x y are calculated as,
R ˙ x y = V t cos ( χ t λ x y ) V u cos ( χ u λ x y ) ,
R x y λ ˙ x y = V t sin ( χ t λ x y ) V u sin ( χ u λ x y ) .
Figure A1. Planar engagement geometry of UAV target.
Figure A1. Planar engagement geometry of UAV target.
Sensors 22 01116 g0a1
The guidance command, denoted as a x y is the lateral acceleration applied normal to the velocity vector. The guidance command is expressed as,
χ ˙ u = a x y / V u .
The principle of PP is to align χ u towards λ x y [53], so that the UAV velocity vector is directed towards the target. A practical implementation of the pure pursuit uses a feedback law as,
a x y = K a ( χ u λ x y ) , K a > 0 ,
where K a is the gain. However, for PP to be effective, it is important to maintain χ ˙ u = λ ˙ x y , so that the vehicle can point towards the target at all times. Combining Equations (A3) and (A4), the guidance command for simple pure pursuit is given as,
a x y = V u λ ˙ x y K a ( χ u λ x y ) .
The simplicity of implementation and this inherent property of PP guidance to point at the target at all times makes it an ideal choice for vision-based landing applications, as it ensures persistent target detection and tracking as the vehicle closes in for landing.

Appendix A.2. 3D Pure Pursuit Guidance

We extend the 2D pursuit guidance principle to 3D for landing. We consider the target moving on X Y horizontal plane. Assume that the initial location of the UAV is ( x 0 , y 0 , z 0 ) and the target is visible from this initial location. Figure A2 shows the engagement geometry for a UAV landing on the target in a 3D scenario. As the landing is a 3D process, we need the UAV course angle χ u and flight path angle γ information. Expressions for UAV target distance, LOS angle λ , and projection of LOS angle in X Y plane λ x y are given as,
R = ( x t x u ) 2 + ( y t y u ) 2 + ( z t z u ) 2 ,
R x y = ( x t x u ) 2 + ( y t y u ) 2 ,
R z = z t z u ,
λ x y = tan 1 ( y t y u / x t x u ) ,
λ = tan 1 ( z t z u / R x y ) .
Expressions of R ˙ x y , R z ˙ , λ ˙ x y are derived similar to the 2D case.
R ˙ x y = V t cos ( χ t λ x y ) V u cos γ cos ( χ u λ x y ) ,
R ˙ z = V u sin γ ,
R x y λ ˙ x y = V t sin ( χ t λ x y ) V u cos γ sin ( χ u λ x y ) .
Figure A2. UAV target engagement geometry for 3D landing process.
Figure A2. UAV target engagement geometry for 3D landing process.
Sensors 22 01116 g0a2
Guidance command in 3D is given by,
a x y = V u cos γ λ ˙ x y K a ( χ u λ x y ) ,
a z = V u λ ˙ K a ( γ λ ) ,
χ ˙ u = a x y / V u cos γ ,
γ ˙ = a z / V u .
We assume the target is nonmaneuvering and traveling with a constant velocity and heading. The kinematic model for the target and UAV are given as
x t t + Δ t = x t t + V t cos χ t Δ t ,
y t t + Δ t = y t t + V t sin χ t Δ t ,
x u t + Δ t = x u t + V u cos γ cos χ u Δ t ,
y u t + Δ t = y u t + V u cos γ sin χ u Δ t ,
z u t + Δ t = z u t + V u sin γ Δ t .

Appendix B. Proof: Convergence of Closing Velocity to Zero Using Log Polynomial Controller

Theorem A1.
Let f(x) be a log of polynomial function of the ratio of current to initial line of sight distance ( R x y / R x y 0 ). Then, the closing velocity ( R ˙ x y ) asymptotically decreases to zero as the distance goes to zero if V u = f ( x ) V t .
Proof. 
In PP, the aim is to continuously point towards the target and the guidance command is aimed at maintaining χ u = λ x y .
Therefore Equation (A1) reduces to
R ˙ x y = V t cos ( χ t λ x y ) V u .
As, the UAV approaches the target, R x y → 0. In PP as R x y 0 , it becomes a tail chase scenario where the UAV is just behind the target and χ u χ t . But since PP inherently maintains χ u λ x y it can be deduced that as R x y 0 , λ x y χ t , reducing Equation (A23) to
lim R x y 0 R ˙ x y = V t V u .
The variation in velocity is given as,
V u = V t f ( x ) ,
where,
N = V u 0 / V t 0 ,
x = R x y / R x y 0 .
Initially, R x y = R x y 0 , hence x = 1. Therefore, from Equations (5) and (A25)
log n ( 1 + a x + b x 2 + c x 3 ) = 1 ,
V u = N V t i f R x y = R x y 0 .
As UAV approaches the target and R x y → 0, x→ 0 and l o g n ( 1 + a x + b x 2 + c x 3 ) → 0. As a result,
lim R x y 0 V u V t , and
lim R x y 0 R ˙ x y 0 .

References

  1. Geng, L.; Zhang, Y.F.; Wang, J.J.; Fuh, J.Y.H.; Teo, S.H. Mission planning of autonomous UAVs for urban surveillance with evolutionary algorithms. In Proceedings of the IEEE International Conference on Control and Automation, Hangzhou, China, 12–14 June 2013; pp. 828–833. [Google Scholar]
  2. Tokekar, P.; Hook, J.V.; Mulla, D.; Isler, V. Sensor planning for a symbiotic UAV and UGV system for precision agriculture. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Tokyo, Japan, 3–7 November 2013; pp. 5321–5326. [Google Scholar]
  3. Herwitz, S.; Johnson, L.; Dunagan, S.; Higgins, R.; Sullivan, D.; Zheng, J.; Lobitz, B.; Leung, J.; Gallmeyer, B.; Aoyagi, M.; et al. Imaging from an unmanned aerial vehicle: Agricultural surveillance and decision support. Comput. Electron. Agric. 2004, 44, 49–61. [Google Scholar] [CrossRef]
  4. Waharte, S.; Trigoni, N. Supporting Search and Rescue Operations with UAVs. In Proceedings of the International Conference on Emerging Security Technologies, Canterbury, UK, 6–7 September 2010; pp. 142–147. [Google Scholar]
  5. Cesetti, A.; Frontoni, E.; Mancini, A.; Zingaretti, P.; Longhi, S. A Vision-Based Guidance System for UAV Navigation and Safe Landing using Natural Landmarks. In International Symposium on UAVs; Valavanis, K.P., Beard, R., Oh, P., Ollero, A., Piegl, L.A., Shim, H., Eds.; Springer: Dordrecht, The Netherlands, 2010; pp. 233–257. [Google Scholar]
  6. Skulstad, R.; Syversen, C.; Merz, M.; Sokolova, N.; Fossen, T.; Johansen, T. Autonomous net recovery of fixed-wing UAV with single-frequency carrier-phase differential GNSS. IEEE Aerosp. Electron. Syst. Mag. 2015, 30, 18–27. [Google Scholar] [CrossRef]
  7. Falanga, D.; Zanchettin, A.; Simovic, A.; Delmerico, J.; Scaramuzza, D. Vision-based autonomous quadrotor landing on a moving platform. In Proceedings of the IEEE International Symposium on Safety, Security and Rescue Robotics, Shanghai, China, 11–13 October 2017; pp. 11–13. [Google Scholar]
  8. Wang, L.; Bai, X. Quadrotor Autonomous Approaching and Landing on a Vessel Deck. J. Intell. Robot. Syst. 2018, 92, 125–143. [Google Scholar] [CrossRef]
  9. Jin, R.; Owais, H.M.; Lin, D.; Song, T.; Yuan, Y. Ellipse proposal and convolutional neural network discriminant for autonomous landing marker detection. J. Field Robot. 2019, 36, 6–16. [Google Scholar] [CrossRef] [Green Version]
  10. Arora, S.; Jain, S.; Scherer, S.; Nuske, S.; Chamberlain, L.; Singh, S. Infrastructure-free shipdeck tracking for autonomous landing. In Proceedings of the IEEE International Conference on Robotics and Automation, Karlsruhe, Germany, 6–10 May 2013; pp. 323–330. [Google Scholar]
  11. Iwakura, D.; Wang, W.; Nonami, K.; Haley, M. Movable Range-Finding Sensor System and Precise Automated Landing of Quad-Rotor MAV. J. Syst. Des. Dyn. 2011, 5, 17–29. [Google Scholar] [CrossRef] [Green Version]
  12. Herissé, B.; Hamel, T.; Mahony, R.; Russotto, F.X. Landing a VTOL Unmanned Aerial Vehicle on a Moving Platform Using Optical Flow. IEEE Trans. Robot. 2012, 28, 77–89. [Google Scholar] [CrossRef]
  13. Jung, Y.; Bang, H.; Lee, D. Robust marker tracking algorithm for precise UAV vision-based autonomous landing. In Proceedings of the International Conference on Control, Automation and Systems, Busan, Korea, 13–16 October 2015; pp. 443–446. [Google Scholar]
  14. Saripalli, S.; Montgomery, J.F.; Sukhatme, G.S. Visually guided landing of an unmanned aerial vehicle. IEEE Trans. Robot. Autom. 2003, 19, 371–380. [Google Scholar] [CrossRef] [Green Version]
  15. Lange, S.; Sunderhauf, N.; Protzel, P. A vision-based onboard approach for landing and position control of an autonomous multirotor UAV in GPS-denied environments. In Proceedings of the IEEE International Conference on Advanced Robotics, Munich, Germany, 22–26 June 2009; pp. 1–6. [Google Scholar]
  16. Hoang, T.; Bayasgalan, E.; Wang, Z.; Tsechpenakis, G.; Panagou, D. Vision-based target tracking and autonomous landing of a quadrotor on a ground vehicle. In Proceedings of the American Control Conference (ACC), Seattle, WA, USA, 24–26 May 2017; pp. 5580–5585. [Google Scholar]
  17. Abu-Jbara, K.; Alheadary, W.; Sundaramorthi, G.; Claudel, C. A robust vision-based runway detection and tracking algorithm for automatic UAV landing. In Proceedings of the International Conference on Unmanned Aircraft Systems, Denver, CO, USA, 9–12 June 2015; pp. 1148–1157. [Google Scholar]
  18. Ghamry, K.A.; Dong, Y.; Kamel, M.A.; Zhang, Y. Real-time autonomous take-off, tracking and landing of UAV on a moving UGV platform. In Proceedings of the 2016 24th Mediterranean Conference on Control and Automation (MED), Athens, Greece, 21–24 June 2016; pp. 1236–1241. [Google Scholar]
  19. Daly, J.M.; Ma, Y.; Waslander, S.L. Coordinated landing of a quadrotor on a skid-steered ground vehicle in the presence of time delays. Auton. Robot. 2015, 38, 179–191. [Google Scholar] [CrossRef]
  20. Li, Z.; Meng, C.; Zhou, F.; Ding, X.; Wang, X.; Zhang, H.; Guo, P.; Meng, X. Fast vision-based autonomous detection of moving cooperative target for unmanned aerial vehicle landing. J. Field Robot. 2019, 36, 34–48. [Google Scholar] [CrossRef] [Green Version]
  21. Serra, P.; Cunha, R.; Hamel, T.; Cabecinhas, D.; Silvestre, C. Landing of a quadrotor on a moving target using dynamic image-based visual servo control. IEEE Trans. Robot. 2016, 32, 1524–1535. [Google Scholar] [CrossRef]
  22. Lee, D.; Ryan, T.; Kim, H.J. Autonomous landing of a VTOL UAV on a moving platform using image-based visual servoing. In Proceedings of the 2012 IEEE International Conference on Robotics and Automation, Saint Paul, MN, USA, 14–18 May 2012; pp. 971–976. [Google Scholar]
  23. Wynn, J.S.; McLain, T.W. Visual servoing with feed-forward for precision shipboard landing of an autonomous multirotor. In Proceedings of the 2019 American Control Conference (ACC), Philadelphia, PA, USA, 10–12 July 2019; pp. 3928–3935. [Google Scholar]
  24. Rodriguez-Ramos, A.; Sampedro, C.; Bavle, H.; Milosevic, Z.; Garcia-Vaquero, A.; Campoy, P. Towards fully autonomous landing on moving platforms for rotary unmanned aerial vehicles. In Proceedings of the 2017 International Conference on Unmanned Aircraft Systems (ICUAS), Miami, FL, USA, 13–16 June 2017; pp. 170–178. [Google Scholar]
  25. Vlantis, P.; Marantos, P.; Bechlioulis, C.P.; Kyriakopoulos, K.J. Quadrotor landing on an inclined platform of a moving ground vehicle. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), Seattle, WA, USA, 26–30 May 2015; pp. 2202–2207. [Google Scholar]
  26. Baca, T.; Stepan, P.; Spurny, V.; Hert, D.; Penicka, R.; Saska, M.; Thomas, J.; Loianno, G.; Kumar, V. Autonomous landing on a moving vehicle with an unmanned aerial vehicle. J. Field Robot. 2019, 36, 874–891. [Google Scholar] [CrossRef]
  27. Borowczyk, A.; Nguyen, D.T.; Nguyen, A.P.V.; Nguyen, D.Q.; Saussié, D.; Le Ny, J. Autonomous landing of a quadcopter on a high-speed ground vehicle. J. Guid. Control. Dyn. 2017, 40, 2378–2385. [Google Scholar] [CrossRef] [Green Version]
  28. Gautam, A.; Sujit, P.; Saripalli, S. A survey of autonomous landing techniques for UAVs. In Proceedings of the International Conference on Unmanned Aircraft Systems, Orlando, FL, USA, 27–30 May 2014; pp. 1210–1218. [Google Scholar] [CrossRef]
  29. Voos, H. Nonlinear control of a quadrotor micro-UAV using feedback-linearization. In Proceedings of the IEEE International Conference on Mechatronics, Málaga, Spain, 14–17 April 2009; pp. 1–6. [Google Scholar]
  30. Ahmed, B.; Pota, H.R. Backstepping-based landing control of a RUAV using tether incorporating flapping correction dynamics. In Proceedings of the American Control Conference, Seattle, WA, USA, 11–13 June 2008; pp. 2728–2733. [Google Scholar]
  31. Shue, S.P.; Agarwal, R.K. Design of Automatic Landing Systems Using Mixed H2/H Control. J. Guid. Control. Dyn. 1999, 22, 103–114. [Google Scholar] [CrossRef]
  32. Malaek, S.M.B.; Sadati, N.; Izadi, H.; Pakmehr, M. Intelligent autolanding controller design using neural networks and fuzzy logic. In Proceedings of the Asian Control Conference, Melbourne, Australia, 20–23 July 2004; Volume 1, pp. 365–373. [Google Scholar]
  33. Nho, K.; Agarwal, R.K. Automatic landing system design using fuzzy logic. J. Guid. Control. Dyn. 2000, 23, 298–304. [Google Scholar] [CrossRef]
  34. Mori, R.; Suzuki, S.; Sakamoto, Y.; Takahara, H. Analysis of visual cues during landing phase by using neural network modeling. J. Aircr. 2007, 44, 2006–2011. [Google Scholar] [CrossRef]
  35. Ghommam, J.; Saad, M. Autonomous landing of a quadrotor on a moving platform. IEEE Trans. Aerosp. Electron. Syst. 2017, 53, 1504–1519. [Google Scholar] [CrossRef]
  36. Gautam, A.; Sujit, P.; Saripalli, S. Application of guidance laws to quadrotor landing. In Proceedings of the 2015 International Conference on Unmanned Aircraft Systems (ICUAS), Denver, CO, USA, 9–12 June 2015; pp. 372–379. [Google Scholar]
  37. Beard, R.W.; McLain, T.W. Small Unmanned Aircraft: Theory and Practice; Princeton University Press: Princeton, NJ, USA, 2012. [Google Scholar]
  38. Kim, H.J.; Kim, M.; Lim, H.; Park, C.; Yoon, S.; Lee, D.; Choi, H.; Oh, G.; Park, J.; Kim, Y. Fully Autonomous Vision-Based Net-Recovery Landing System for a Fixed-Wing UAV. IEEE/ASME Trans. Mechatronics 2013, 18, 1320–1333. [Google Scholar] [CrossRef]
  39. Yoon, S.; Jin Kim, H.; Kim, Y. Spiral Landing Trajectory and Pursuit Guidance Law Design for Vision-Based Net-Recovery UAV. In Proceedings of the AIAA Guidance, Navigation, and Control Conference and Exhibit, Chicago, IL, USA, 10–13 August 2009. [Google Scholar] [CrossRef]
  40. Byoung-Mun, M.; Min-Jea, T.; David Hyunchul, S.; HyoChoong, B. Guidance law for vision-based automatic landing of UAV. Int. J. Aeronaut. Space Sci. 2007, 8, 46–53. [Google Scholar]
  41. Barber, D.B.; Griffiths, S.R.; McLain, T.W.; Beard, R.W. Autonomous landing of miniature aerial vehicles. J. Aerosp. Comput. Inform. Commun. 2007, 4, 770–784. [Google Scholar] [CrossRef]
  42. Huh, S.; Shim, D.H. A Vision-Based Automatic Landing Method for Fixed-Wing UAVs. In Selected Papers from the 2nd International Symposium on UAVs, Reno, Nevada, USA, 8–10 June 2009; Valavanis, K.P., Beard, R.O.P.O.A., Piegl, L.A., Shim, H., Eds.; Springer: Dordrecht, The Netherlands, 2010; pp. 217–231. [Google Scholar]
  43. Gautam, A.; Sujit, P.; Saripalli, S. Autonomous Quadrotor Landing Using Vision and Pursuit Guidance. IFAC-PapersOnLine 2017, 50, 10501–10506. [Google Scholar] [CrossRef]
  44. Gautam, A.; Sujit, P.; Saripalli, S. Vision Based Robust Autonomous Landing of a Quadrotor on a Moving Target. In International Symposium on Experimental Robotics; Springer: Berlin/Heidelberg, Germany, 2018; pp. 96–105. [Google Scholar]
  45. Gautam, A.; Ratnoo, A.; Sujit, P. Log Polynomial Velocity Profile for Vertical Landing. J. Guid. Control. Dyn. 2018, 41, 1617–1623. [Google Scholar] [CrossRef]
  46. Olson, E. AprilTag: A robust and flexible visual fiducial system. In Proceedings of the IEEE International Conference on Robotics and Automation, Shanghai, China, 9–13 May 2011; pp. 3400–3407. [Google Scholar]
  47. Shah, S.; Dey, D.; Lovett, C.; Kapoor, A. AirSim: High-Fidelity Visual and Physical Simulation for Autonomous Vehicles. arXiv 2017, arXiv:1705.05065. [Google Scholar]
  48. Games, E. Unreal Engine. 2007. Available online: https://www.unrealengine.com (accessed on 27 December 2021).
  49. Quigley, M.; Conley, K.; Gerkey, B.; Faust, J.; Foote, T.; Leibs, J.; Wheeler, R.; Ng, A.Y. ROS: An open-source Robot Operating System. ICRA Workshop Open Source Softw. 2009, 3, 5. [Google Scholar]
  50. Bradski, G. The OpenCV Library. Dr. Dobb’S J. Softw. Tools 2000, 25, 120–123. [Google Scholar]
  51. Lee, H.; Jung, S.; Shim, D.H. Vision-based UAV landing on the moving vehicle. In Proceedings of the 2016 International conference on unmanned aircraft systems (ICUAS), Arlington, VA, USA, 7–10 June 2016; pp. 1–7. [Google Scholar]
  52. Chang, C.W.; Lo, L.Y.; Cheung, H.C.; Feng, Y.; Yang, A.S.; Wen, C.Y.; Zhou, W. Proactive Guidance for Accurate UAV Landing on a Dynamic Platform: A Visual–Inertial Approach. Sensors 2022, 22, 404. [Google Scholar] [CrossRef] [PubMed]
  53. Shneydor, N.A. Missile Guidance and Pursuit: Kinematics, Dynamics and Control; Elsevier: Amsterdam, The Netherlands, 1998. [Google Scholar]
Figure 1. Block diagram of a vision-assisted guidance landing system.
Figure 1. Block diagram of a vision-assisted guidance landing system.
Sensors 22 01116 g001
Figure 2. (a) Target setup for autonomous landing. Inset figure shows extracted contour of red blob (b) One of detected AprilTags.
Figure 2. (a) Target setup for autonomous landing. Inset figure shows extracted contour of red blob (b) One of detected AprilTags.
Sensors 22 01116 g002
Figure 3. Effect of variation of ( a , b ) on closing velocity profile for moving target landing; V u 0 = 2 m/s, V t = 1 m/s and n = 10 .
Figure 3. Effect of variation of ( a , b ) on closing velocity profile for moving target landing; V u 0 = 2 m/s, V t = 1 m/s and n = 10 .
Sensors 22 01116 g003
Figure 4. A snapshot of simulation setup consisting of airsim flying quadrotor vehicle and a ground vehicle as landing target.
Figure 4. A snapshot of simulation setup consisting of airsim flying quadrotor vehicle and a ground vehicle as landing target.
Sensors 22 01116 g004
Figure 5. Landing trajectory followed by unmanned aerial vehicle (UAV) when ground vehicle moves in a near straight line trajectory at (a) V t = 1 m/s (b) V t = 2 m/s (c) V t = 3 m/s .
Figure 5. Landing trajectory followed by unmanned aerial vehicle (UAV) when ground vehicle moves in a near straight line trajectory at (a) V t = 1 m/s (b) V t = 2 m/s (c) V t = 3 m/s .
Sensors 22 01116 g005
Figure 6. (a) Closing velocity profile ( x y plane) of UAV when vehicle moves in a near straight line trajectory at 2 m/s. (b) Descend velocity of UAV when ground vehicle moves in a near straight line trajectory at 2 m/s.
Figure 6. (a) Closing velocity profile ( x y plane) of UAV when vehicle moves in a near straight line trajectory at 2 m/s. (b) Descend velocity of UAV when ground vehicle moves in a near straight line trajectory at 2 m/s.
Sensors 22 01116 g006
Figure 7. (a) Landing trajectory followed by UAV when ground vehicle moves along a circular path (15 m radius); (b) closing velocity profile ( x y plane) of the UAV when ground vehicle moves along a circular path (15 m radius); (c) descend velocity of the UAV when ground vehicle moves along a circular path (15 m radius).
Figure 7. (a) Landing trajectory followed by UAV when ground vehicle moves along a circular path (15 m radius); (b) closing velocity profile ( x y plane) of the UAV when ground vehicle moves along a circular path (15 m radius); (c) descend velocity of the UAV when ground vehicle moves along a circular path (15 m radius).
Sensors 22 01116 g007
Figure 8. Landing trajectory followed by UAV when ground vehicle moves along a circular path (5 m radius).
Figure 8. Landing trajectory followed by UAV when ground vehicle moves along a circular path (5 m radius).
Sensors 22 01116 g008
Figure 9. (a) Kalman filter performance for estimating target information (b) Trajectory followed by UAV when using estimated target information.
Figure 9. (a) Kalman filter performance for estimating target information (b) Trajectory followed by UAV when using estimated target information.
Sensors 22 01116 g009
Figure 10. (a) Hardware setup and information flow diagram consisting of 3DR Solo (quadrotor), Odroid U3+, and PtGrey USB camera; (b) quadcopter with whole hardware interface about to land on target. Inset Figures show 3DR solo setup and detected target from onboard camera.
Figure 10. (a) Hardware setup and information flow diagram consisting of 3DR Solo (quadrotor), Odroid U3+, and PtGrey USB camera; (b) quadcopter with whole hardware interface about to land on target. Inset Figures show 3DR solo setup and detected target from onboard camera.
Sensors 22 01116 g010
Figure 11. Vision and control pipeline for autonomous landing.
Figure 11. Vision and control pipeline for autonomous landing.
Sensors 22 01116 g011
Figure 12. Trial 1: trajectory and velocity profile characteristics for moving target landing. (a) Trajectory followed by the quadcopter to land on a moving target; (b) lateral velocity profile ( a , b , c ) = ( 10 , 130 , 131 ) ; (c) longitudinal velocity profile ( a , b , c ) = ( 40 , 130 , 161 ) (d) V z profile ( a , b ) = ( 6 , 3 ) .
Figure 12. Trial 1: trajectory and velocity profile characteristics for moving target landing. (a) Trajectory followed by the quadcopter to land on a moving target; (b) lateral velocity profile ( a , b , c ) = ( 10 , 130 , 131 ) ; (c) longitudinal velocity profile ( a , b , c ) = ( 40 , 130 , 161 ) (d) V z profile ( a , b ) = ( 6 , 3 ) .
Sensors 22 01116 g012
Figure 13. Straight line moving target trial 2: (a) vehicle’s trajectory; (b) longitudinal velocity profile for landing on a straight-line moving target for a different set of log polynomial parameters ( a , b , c ) = ( 55 , 115 , 161 ) .
Figure 13. Straight line moving target trial 2: (a) vehicle’s trajectory; (b) longitudinal velocity profile for landing on a straight-line moving target for a different set of log polynomial parameters ( a , b , c ) = ( 55 , 115 , 161 ) .
Sensors 22 01116 g013
Figure 14. Landing sequence on a ground moving target.
Figure 14. Landing sequence on a ground moving target.
Sensors 22 01116 g014
Figure 15. Case1: (a) UAV-target trajectory for a nonstraight-line moving target landing; (b) lateral velocity profile of UAV ( a , b , c ) = ( 40 , 130 , 161 ) ; (c) longitudinal velocity profile of UAV ( a , b , c ) = ( 40 , 130 , 161 ) (d) z velocity profile of the UAV ( a , b ) = ( 6 , 3 ) .
Figure 15. Case1: (a) UAV-target trajectory for a nonstraight-line moving target landing; (b) lateral velocity profile of UAV ( a , b , c ) = ( 40 , 130 , 161 ) ; (c) longitudinal velocity profile of UAV ( a , b , c ) = ( 40 , 130 , 161 ) (d) z velocity profile of the UAV ( a , b ) = ( 6 , 3 ) .
Sensors 22 01116 g015
Figure 16. Case2: UAV-target trajectory for a nonstraight-line moving target landing.
Figure 16. Case2: UAV-target trajectory for a nonstraight-line moving target landing.
Sensors 22 01116 g016
Figure 17. Moving target occlusion results. (a) Target detected and tracking flag; (b) trajectory followed by UAV to land on target; (c) altitude profile of UAV; (d) V z profile of UAV.
Figure 17. Moving target occlusion results. (a) Target detected and tracking flag; (b) trajectory followed by UAV to land on target; (c) altitude profile of UAV; (d) V z profile of UAV.
Sensors 22 01116 g017
Figure 18. (a) Box plot of successful experiment trials for different target trajectory cases. (b) Depiction of vehicle landing position in different experiment runs.
Figure 18. (a) Box plot of successful experiment trials for different target trajectory cases. (b) Depiction of vehicle landing position in different experiment runs.
Sensors 22 01116 g018
Figure 19. A depiction of a quadrotor trying to land on a moving target with velocity V t .
Figure 19. A depiction of a quadrotor trying to land on a moving target with velocity V t .
Sensors 22 01116 g019
Table 1. Categorization of experimental trials according to target trajectory and target occlusion.
Table 1. Categorization of experimental trials according to target trajectory and target occlusion.
Target TrajectoryNo. of
Trials
Final Vehicle Distance from Target
Center for Successful Trials (in M)
Unsuccessful Landings
(Failure Cases)
Best Landing
Accuracy
Average Landing
Time
Straight line trajectory,
no occlusion
97 trials 0.20 , 2 trials 0.35 None 0.035  m56.5 s
Straight line trajectory,
with occlusion
42 trials 0.20 , 1 trial 0.35 1 with an absolute error
of 0.47  m
0.11  m1 min 2 s
Random trajectory, no
occlusion
94 trials 0.20 , 3 trials 0.35 2 with absolute errors
0.45  m and 0.54  m
respectively
0.055  m1 min 38 s
Random trajectory,
with occlusion
53 trials 0.35 2 with absolute errors
0.60  m 0.55  m
respectively
0.25  m1min 37 s
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Gautam, A.; Singh, M.; Sujit, P.B.; Saripalli, S. Autonomous Quadcopter Landing on a Moving Target. Sensors 2022, 22, 1116. https://doi.org/10.3390/s22031116

AMA Style

Gautam A, Singh M, Sujit PB, Saripalli S. Autonomous Quadcopter Landing on a Moving Target. Sensors. 2022; 22(3):1116. https://doi.org/10.3390/s22031116

Chicago/Turabian Style

Gautam, Alvika, Mandeep Singh, Pedda Baliyarasimhuni Sujit, and Srikanth Saripalli. 2022. "Autonomous Quadcopter Landing on a Moving Target" Sensors 22, no. 3: 1116. https://doi.org/10.3390/s22031116

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop