Reinforcement Learning for Planning and Scheduling in Aviation

187627-Thumbnail Image.png
Description
Aviation is a complicated field that involves a wide range of operations, from commercial airline flights to Unmanned Aerial Systems (UAS). Planning and scheduling are essential components in the aviation industry that play a significant role in ensuring safe and

Aviation is a complicated field that involves a wide range of operations, from commercial airline flights to Unmanned Aerial Systems (UAS). Planning and scheduling are essential components in the aviation industry that play a significant role in ensuring safe and efficient operations. Reinforcement Learning (RL) has received increasing attention in recent years due to its capability to enable autonomous decision-making. To investigate the potential advantages and effectiveness of RL in aviation planning and scheduling, three topics are explored in-depth, including obstacle avoidance, task-oriented path planning, and maintenance scheduling. A dynamic and probabilistic airspace reservation concept, called Dynamic Anisotropic (DA) bound, is first developed for UAS, which can be added around the UAS as the separation requirement. A model based on Q-leaning is proposed to integrate DA bound with path planning for obstacle avoidance. Moreover, A deep reinforcement learning algorithm based on Proximal Policy Optimization (PPO) is proposed to guide the UAS to destinations while avoiding obstacles through continuous control. Results from case studies demonstrate that the proposed model can provide accurate and robust guidance and resolve conflict with a success rate of over 99%. Next, the single-UAS path planning problem is extended to a multi-agent system where agents aim to accomplish their own complex tasks. These tasks involve non-Markovian reward functions and can be specified using reward machines. Both cooperative and competitive environments are explored. Decentralized Graph-based reinforcement learning using Reward Machines (DGRM) is proposed to improve computational efficiency for maximizing the global reward in a graph-based Markov Decision Process (MDP). Q-learning with Reward Machines for Stochastic Games (QRM-SG) is developed to learn the best-response strategy for each agent in a competitive environment. Furthermore, maintenance scheduling is investigated. The purpose is to minimize the system maintenance cost while ensuring compliance with reliability requirements. Maintenance scheduling is formulated as an MDP and determines when and what maintenance operations to conduct. A Linear Programming-enhanced RollouT (LPRT) method is developed to solve both constrained deterministic and stochastic maintenance scheduling with an infinite horizon. LPRT categorizes components according to their health condition and makes decisions for each category.
Date Created
2023
Agent

Numerical Modeling of Cake Formation and Permeate Flux Decline in Membrane Filtration Using OpenFOAM

156167-Thumbnail Image.png
Description
Membrane filtration is an important technology in industry. In the past few decades, equations have been developed from experimental results to predict cake formation and permeate flux decline in the membrane filtration process. In the current work, the block of

Membrane filtration is an important technology in industry. In the past few decades, equations have been developed from experimental results to predict cake formation and permeate flux decline in the membrane filtration process. In the current work, the block of particles on membrane surface is achieved by setting surface flux on membrane surface zero. This approach is implemented for both microfiltration and nanofiltration using OpenFOAM. Moreover, a new method to deal with cake resistance for nanofiltration is introduced. Cake resistance is applied to both cake and membrane. To validate the new techniques, results of crossflow microfiltration are compared to theoretical results and results of two crossflow nanofiltration cases are compared to experimental data. In addition, the new techniques are applied to dead end filtration to observe the different structure of the cake and explore the effect of resistance on velocity profile.
Date Created
2018
Agent