Loading...
Thumbnail Image
Publication

Bio-inspired Reinforcement Learning & Predictive Flight Controllers for Unmanned Aerial Systems

Chowdhury, Md Mozammal Hosain
Citations
Altmetric:
Abstract
Recent advancements in computer technologies have dramatically increased the onboard processing power of autonomous aircraft as well as the performance of autopilot systems. The combination of exponential growth in applications of autonomous aircraft and computationally potent avionic systems provides opportunities and demand for adaptive, learning, and cognitive flight controller methods. This work investigates two flight controller methods: model predictive control (MPC) and reinforcement learning (RL) for fixed-wing UASs.In addition to the adaptivity of flight controllers, the transferability of controllers between aircraft in the same class is highly desirable. Therefore, using and adapting a flight controller between different aircraft can save overhead costs, time, and effort. This work uniquely presents a method to develop model-agnostic RL-based flight controllers capable of controlling aircraft platforms in the same class/category but different dynamic models (e.g., fixed-wing twin-boom pusher UASs weighing less than 12 lbs (5.4 kg)). The proposed method uses dynamic randomization of aircraft stability and control derivatives to develop the training environment and incorporate the memory functions into the policy using a recurrent neural network (RNN). Flying autonomous aircraft in constraint spaces (e.g., metropolitan areas, etc.) can result in a phase shift in the control signal and undesirable and sustained oscillations. Therefore, a unified RL-based longitudinal control policy is also developed to mitigate the oscillation issues due to coupling between outer (guidance) and inner loop control blocks. In addition, the dynamic model of aircraft is improved using the cross-entropy method (CEM) algorithm. CEM is a derivative-free optimization algorithm that supports process parallelization and can learn from actual flight test data to improve the fidelity of the dynamic model of aircraft. The MPC is formulated as a receding horizon control problem for the linearized aircraft model with a set of control constraints which is solved using the Sequential Quadratic Programming (SQP) optimization method. Uniquely, the stability of the MPC closed-loop system (GNC) is analyzed through Lyapunov theory, assessed by 1000 Monte Carlo numerical simulations. All the developed flight controllers were uniquely verified and validated (V&V) in actual flight tests and compared with two base flight controllers (commercial-off-the-shelf and LQR) explicitly designed for the testbed platforms. Two UAS platforms with different dynamics and weather conditions from low to medium wind intensities were chosen to evaluate proposed methods for practically viable flight controllers. MPC flight controllers' sensitivity to pertinent design parameters, such as controller update rate and the prediction horizon, were evaluated through flight tests.Finally, the controllers' performance, stability, and resilience to the partial failure of the control surfaces were investigated and documented.
Description
Date
2023-05-31
Journal Title
Journal ISSN
Volume Title
Publisher
University of Kansas
Collections
Research Projects
Organizational Units
Journal Issue
Keywords
Aerospace engineering, Artificial intelligence, Computer science, Artificial Intelligence, Autonomous Flight Control, Model agnostic, Model Predictive Control, Reinforcement Learning, Unmanned Aerial System
Citation
DOI
Embedded videos