ML_Masters_Proposal-Midterm




Click here to view the Final Report

Click here to view the Midterm Checkpoint




ML_Masters_Proposal

AerialVLN w Fuel Constraints

By Christine Yewon Kim, Mahika Raj, Devanshi Sood, Joseph C Miao, and Song Yue David Li



Report:


Introduction/Background

Literature

In the midst of Vision and Language Navigation innovation, AerialVLN was proposed, a UAV-based VLN task for outdoor environments [1]. However, as energy awareness is not widely discussed in VLN, the typical path-length objective of existing approaches does not directly minimize energy consumption, nor allows constraining the energy of individual paths by battery capacity [2]. We will utilize a combination of Supervised Learning and Reinforcement Learning to better consider energy.

Dataset Description

This dataset includes AerialVLN and AerialVLN-S annotated images for training and evaluation, a total of 32.4GB.

Dataset Link
AirVLN GitHub Repository

Problem

With the new constraints, specifically fuel capacity and limits, we focus on aerial navigation in the sky, UAV-based and aimed towards outdoor environments.

Motivation

Many existing VLN tasks are built for agents that navigate on the ground, either indoors or outdoors. However, some tasks require intelligent agents to operate in the sky, such as UAV-based goods delivery, traffic/security patrol, and scenery tours [1]. Most importantly, aerial navigation is a field with much work to be done.


Methods

Data Preprocessing Methods

3 data preprocessing methods we are considering are fuel tokenization, state-space normalization, and trajectory-based windowing:

ML Algorithms and Models

We considered 2 supervised learning models: Recurrent Neural Networks and Transformers. These would function by embedding the instructions, visuals, and fuel as inputs, then outputting a low-cost trajectory.

Alternatively, and perhaps more intuitively, we can pursue Proximal Policy Optimization (PPO), a reinforcement learning algorithm. Here, we would use fuel capacity as a Lagrange penalty.


Learning Methods

Supervised Training

RL Method


Potential Results and Discussion

We will evaluate the model using success rate, path efficiency, mean fuel consumption, and constraint violation rate.

We aim to create a model that produces smoother paths by reducing sharp turns, lowering fuel use, and improving success rate under energy constraints.

Expected results are that the proposed supervised and reinforcement learning approaches will produce more efficient flight paths, lower average fuel consumption, and better respect energy limits, while still completing navigation tasks at a strong rate.


Citations

[1]
Liu et al. (2023): AerialVLN: Vision-and-Language Navigation for UAVs (The primary AirVLN paper).

[2]
Pereira et al. (2025): Energy-Aware Coverage Path Planner for Multirotor UAVs (For the fuel modeling aspect).

[3]
Fabio Morbidi, Roel Cano, David Lara. Minimum-Energy Path Generation for a Quadrotor UAV. IEEE International Conference on Robotics and Automation, May 2016, Stockholm, Sweden. Ffhal01276199v2f

Other References




Gantt Chart

Project Image




Contribution Table

Project Image