The BlueRov2 platform at the Flinders University pool facility at Tonsley was used for the BIER testing experiments. (Image: Courtesy of the researchers)

Unmanned Underwater Vehicles (UUVs) are used around the world to conduct difficult environmental, remote, oceanic, defense, and rescue missions in often unpredictable and harsh conditions.

A new study led by Flinders University and French researchers has used a novel bio-inspired computing artificial intelligence solution to improve the potential of UUVs and other adaptive control systems to operate more reliability in rough seas and other unpredictable conditions.

This innovative approach, using the Biologically-Inspired Experience Replay (BIER) method, has been published by the Institute of Electrical and Electronics Engineers journal IEEE Access.

Unlike conventional methods, BIER aims to overcome data inefficiency and performance degradation by leveraging incomplete but valuable recent experiences, explained first author Dr. Thomas Chaffre.

“The outcomes of the study demonstrated that BIER surpassed standard Experience Replay methods, achieving optimal performance twice as fast as the latter in the assumed UUV domain. The method showed exceptional adaptability and efficiency, exhibiting its capability to stabilize the UUV in varied and challenging conditions,” said Chaffre.

The method incorporates two memory buffers, one focusing on recent state-action pairs and the other emphasizing positive rewards.

To test the effectiveness of the proposed method, researchers conducted simulated scenarios using a robot operating system (ROS)-based UUV simulator and gradually increasing the scenarios’ complexity. These scenarios varied in target velocity values and the intensity of current disturbances.

Senior author, Flinders University Associate Professor in AI and Robotics, Paulo Santos said the BIER method’s success holds promise for enhancing adaptability and performance in various fields requiring dynamic, adaptive control systems.

UUVs’ capabilities in mapping, imaging, and sensor controls are improving, including with Deep Reinforcement Learning (DRL), which is rapidly advancing the adaptive control responses to underwater disturbances that UUVs can encounter.

However, the efficiency of these methods gets challenged when faced with unforeseen variations in real-world applications. The complex dynamics of the underwater environment limit the observability of UUV maneuvering tasks, making it difficult for existing DRL methods to perform optimally.

The introduction of BIER marks a significant step forward in enhancing the effectiveness of deep reinforcement learning method in general.

Its ability to efficiently navigate uncertain and dynamic environments signifies a promising advancement in the area of adaptive control systems, according to the researchers.

Source