This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Word2Wave: Language Driven Mission Programming for
Efficient Subsea Deployments of Marine Robots

Ruo Chen, David Blow, Adnan Abdullah, and Md Jahidul Islam
Email: {chenruo@,david.blow@,adnanabdullah@,jahid@ece.}ufl.edu
RoboPI Laboratory, Department of ECE, University of Florida, FL 32611, USA.
This pre-print currently in review at ICRA-2025.Info:  https://robopi.ece.ufl.edu/word2wave.html

Abstract. This paper explores the design and development of a language-based interface for dynamic mission programming of autonomous underwater vehicles (AUVs). The proposed ‘Word2Wave’ (W2W) framework enables interactive programming and parameter configuration of AUVs for remote subsea missions. The W2W framework includes: (i) a set of novel language rules and command structures for efficient language-to-mission mapping; (ii) a GPT-based prompt engineering module for training data generation; (iii) a small language model (SLM)-based sequence-to-sequence learning pipeline for mission command generation from human speech or text; and (iv) a novel user interface for 2D mission map visualization and human-machine interfacing. The proposed learning pipeline adapts an SLM named T5-Small that can learn language-to-mission mapping from processed language data effectively, providing robust and efficient performance. In addition to a benchmark evaluation with state-of-the-art, we conduct a user interaction study to demonstrate the effectiveness of W2W over commercial AUV programming interfaces. Across participants, W2W-based programming required less than 10% time for mission programming compared to traditional interfaces; it is deemed to be a simpler and more natural paradigm for subsea mission programming with a usability score of 76.25. W2W opens up promising future research opportunities on hands-free AUV mission programming for efficient subsea deployments.

1 Introduction

Recent advancements in Large Language Models (LLMs) and speech-based human-machine dialogue frameworks are poised to revolutionize robotics by enabling more natural and spontaneous interactions [1, 2]. These systems allow robots to interpret natural human language for more accessible and interactive operation, especially in challenging field robotics applications. In particular, remote deployments of autonomous robots in subsea inspection, surveillance, and search and rescue operations require dynamic mission adjustments on the fly [3]. Seamless mission parameter adaptation and fast AUV deployment routines are essential features for these applications, which the traditional interfaces often fail to ensure [4, 5, 6, 7].

Refer to caption
(a) A W2W-programmed mission and subsea deployment scenario.
Refer to caption
(b) Corresponding mission maps and sample sensory data.
Figure 1: The proposed Word2Wave (W2W) framework offers a novel user interface for real-time AUV mission programming with natural human language. We demonstrate the effectiveness of W2W over traditional interfaces by subsea mission deployments on a NemoSens AUV platform.

The existing subsea robotics technologies offer predefined planners that require manual configuration of mission parameters in a complex software interface [8, 9]. It is extremely challenging and tedious to program complex missions spontaneously, even for skilled technicians, especially on an undulating vessel when time is of the essence. Natural language-based interfaces have the potential to address these limitations by making mission programming more user-friendly and efficient. Recent works have shown how the reasoning capabilities of LLMs can be applied for mission planning and human-robot interaction (HRI) [1, 10] by using deep vision-language models [11, 12], text-to-action paradigms [13, 14], and embodied reasoning pipelines [10, 1]. Contemporary research demonstrates promising results [15, 16, 4] on language-based human-machine interfaces (HMIs) for subsea mission programming as well.

In this paper, we introduce “Word2Wave”, a small language model (SLM)-based framework for real-time AUV programming in subsea missions. It provides an interactive HMI that uses natural human speech patterns to generate subsea mission plans to perform autonomous remote operations. For implementation, we adapt a SLM training pipeline based on the Text-To-Text Transfer Transformer (T5) [17] small model (T5-Small) to parse natural human speech into a sequence of computer interpretable commands. These commands are subsequently converted to a set of waypoints for AUVs to execute.

We design the language rules of Word2Wave (W2W) with seven atomic commands to support a wide range of subsea mission plans, particularly focusing on subsea surveying, mapping, and inspection tasks [18]. These commands are simple and intuitive, yet powerful tools, to program complex missions without using tedious software interfaces of commercial AUVs [19, 20]. W2W includes all basic operations to program widely used mission patterns such as lawnmower, spiral, ripple, and polygonal trajectories at various configurations. With these flexible operations, W2W allows users to program subsea missions using natural language, similar to how they would describe the mission to a human diver. To this end, we designed a GPT [21]-based prompt engineering module [22] for comprehensive training data generation.

The proposed learning pipeline demonstrates a delicate balance between robustness and efficiency, making it ideal for real-time mission programming. Through comprehensive quantitative and qualitative assessments, we demonstrate that SLMs can capture targeted vocabulary from limited data with a rightly adapted learning pipeline and articulated language rules. Specifically, our adapted T5-Small model provides SOTA performance for sequence-to-sequence learning while offering 2×2\times faster inference rate (of 70.570.5 ms) with 85.1%85.1\% fewer parameters than computationally demanding LLMs such as BART-Large [23]. We also investigated other SLM architectures such as MarianMT [24] for benchmark evaluation based on accuracy and computational efficiency.

Moreover, we develop an interactive user interface (UI) for translating the W2W-generated language tokens into 2D mission maps. Unlike traditional HMIs, it adopts a minimalist design intended for futuristic use cases. Specifically, we envision that users will engage in interactive dialogues for formulating and planning subsea missions. While such HMIs are still an open problem, our proposed UI is significantly more efficient and user-friendly – which we validate by a thorough user interaction study with 1515 participants.

From the user study, we find that on average, participants took less than 10%10\% time for mission programming by W2W compared to using traditional interfaces. The participants, especially those with prior experiences of subsea deployments, preferred using W2W as a simpler and more intuitive programming paradigm. They rated W2W with a usability score [25] of 76.2576.25, validating that it induces less cognitive load and requires minimal technical support for novice users. With these features, the proposed W2W framework takes a step forward to our overarching goal of integrating human-machine dialogue for embodied reasoning and hands-free mission programming of marine robots.

2 Background And Related Work

2.1 Language Models for Human-Machine Embodied Reasoning and HRI

Classical language-based systems focus on deterministic human commands for controlling mobile robots [26, 27]. Traditionally, the open-world navigation with visual goals (ViNGs) [28] or visual-inertial navigation (VIN) [29] pipelines have been mostly independent of human-directed language inputs [30]  [31, 32]. In these systems, language or speech inputs are parsed separately as a control input to the ViNG or VIN systems to achieve motion planning [31, 33] and navigation tasks [34, 35].

With the advent of LLMs, contemporary robotics research have focused on leveraging the power of natural language for more interactive human-robot embodied decision-making and shared autonomy [10, 1]. A key advancement is the development of vision-language models (VLMs) [11, 12] for human-machine embodied reasoning. Huang et al. [14] developed an “Inner Monologue” framework that injects continuous sensorimotor feedback into a LLM which prompts as the robot interacts with the environment. In “InstructPix2Pix” [13], Tim et al. combine an LLM and a text-to-image model for image editing and visual question answering. These features can enable robots to understand the visual content and integrate it with relevant linguistic information, as shown by Wu et al. in the “TidyBot” system [36].

While general-purpose LLMs are resource intensive, Small Language Models (SLMs) are often more suited for targeted robotics applications [37]. SLMs in various zero-shot learning pipelines have been fine-tuned for applications such as long-horizon navigation [38], embodied manipulation [39], and trajectory planning [37, 40]. These are emerging technologies and, thus, ongoing developments for more challenging real-world field robotics applications.

2.2 Subsea Mission Programming Interfaces

Leveraging human expertise is critical for configuring subsea mission parameters of mobile robots because fully autonomous mission planning and navigation are challenging underwater [41, 42, 43, 44]. Human-programmed missions enable AUVs to adapt to dynamic mission objectives and deal with environmental challenges in adverse sensing conditions with no GPS or wireless connectivity [45, 46]. Various HRI frameworks [5, 6, 47], telerobotics consoles [48, 49], and language-based interfaces [50, 20] have been developed for mission programming and parameter reconfiguration. For instance, visual languages such as “RoboChat” [5] and “RoboChatGest” [3] use a sequence of symbolic patterns to communicate simple instructions to the robot via AR-Tag markers and hand gestures, respectively. These and other language paradigms [11] are mainly suited for short-term human-robot cooperative missions [51, 52].

For subsea telerobotics, augmented and virtual reality (AR/VR) interfaces integrated on traditional consoles are gaining popularity in recent times [53, 54, 55]. These offer immersive teleop experiences [56, 57, 48] and improve teleoperators’ perception of environmental semantics [55, 58, 59, 60, 61]. Long-term autonomous missions are generally planned offline in terms of a sequence of waypoints following a specific trajectory [58, 4, 62] such as a lawn mower pattern, perimeter following, fixed-altitude spiral/corkscrew patterns, variable-altitude polygons, etc. Yang et al. [15] proposed a LLM-driven OceanChat system for AUV motion planning in HoloOcean simulator [63]. Despite inspiring results in marine simulators [22, 64, 65], the power of language models for subsea mission deployments [16] on real systems are not explored in depth.

3 Word2Wave: Language Design

3.1 Language Commands And Rules

Designing mission programming languages for subsea robots involves unique considerations due to the particular requirements of marine robotics applications. In Word2Wave, our primary objective is to enable spontaneous human-machine interaction by natural language. We also want to ensure that the high-level abstractions in the Word2Wave (W2W) language integrate with the existing industrial interfaces and simulation environments for seamless adaptations.

Table 1: The language commands, parameters, and command structure notations of Word2Wave are shown.
  \cellcolorgray!10Parameters b: bearing;  d: depth;  s: speed
a: altitude;  t: turns;  r: radius
tab: spacing;  dir: direction;  dist: distance
cw, ccw: clockwise, counter clockwise
  \cellcolorgray!10Command \cellcolorgray!10Symbol \cellcolorgray!10Language Structure
  Start/End S/E [S/E: latitude, longitude]
Move Mv [Mv: b, d, s, d/a, d/a(m)]
Track Tr [Tr: dir, tab, end, d/a, d/a(m)]
Adjust Az [Az: d/a, d/a(m)]
Circle Cr [Cr: t, r, cw/ccw, d/a, d/a(m)]
Spiral SP [SP: t, r, cw/ccw, d/a, d/a(m)]
 

As shown in Table 1, we consider 77 language commands in Word2Wave. Their intended use cases are as follows.

  1. 11.

    Start/End (S/E): is intended to start or end a mission at a given latitude and longitude. These coordinates are two input parameters, taken in decimal degrees with respect to true North and West, respectively.

  2. 22.

    Move (Mv): command is designed to move the AUV to a specified distance (meters) at a given bearing (w.r.t. North) and speed (m/s). If/when the speed is not explicitly stated, a default value of 11 m/s is used.

  3. 33.

    Track (Tr): is used to plan a set of parallel lines given a direction orthogonal to the current bearing (same altitude), spacing between each line, and ending distance.

  4. 44.

    Adjust (Az): generates a waypoint to adjust the AUV’s target depth or altitude at its current location.

  5. 55.

    Circle (Cr): creates a waypoint commanding the AUV to circle around a position for a number of turns at a given radius in a counter/clockwise direction.

  6. 66.

    Spiral (Sp): generates a circular pattern that starts at a central point and then expands outwards over a series of turns out to a specific radius. The spiral direction is set to either clockwise or counterclockwise.

Refer to caption
(a) An example of speech being translated into a series of language tokens pertaining to the mission waypoints.
Refer to caption
(b) A mission sketch and corresponding W2W-generated map are visualized for the waypoint sequence shown in (a).
Figure 2: Demonstration of a sample mission programmed using the proposed Word2Wave (W2W) interface.

3.2 Mission Types And Parameter Selection

W2W can generate arbitrary mission patterns with varying complexity using the language commands of Table 1. The only limitations are the maneuver capabilities of the host AUV. We demonstrate a particular mission programming instance in Fig. 2(a). Mission parameters for each movement command are chosen based on commonly used terms to ensure compatibility; we particularly explore four most widely used mission patterns, as shown in Fig. 2(b). These atomic patterns can be further combined to plan multi-phase composite missions for a given scenario.

  1. \bullet

    Lawnmower (also known as boustrophedon) patterns are ideal for surveying large areas over subsea structures. It is not suited for missions requiring intricate movements or targeted actions. These patterns are most commonly used for sonar-based mapping, as they offer even coverage with a simple and efficient route over a large area.

  2. \bullet

    Polygonal routes are best suited for irregular terrain as they provide more precise control over the AUV trajectory. They are more flexible as polygons can be tailored towards specific mission requirements and adapt to local terrains. Hence, they are used for missions involving inspecting known landmarks or targeted waypoints.

  3. \bullet

    Ripple patterns are defined as a series of concentric circles with either or both varying radii and depths. These are better suited when coverage is needed over a specific area and when equal spacing is important when collecting data. They are best suited for missions where sampling in varying depths of the water column is required.

  4. \bullet

    Spiral paths allow for more concentrated coverage over a specific area. It operates similarly to the ripple pattern but allows for a smooth, continuous trajectory that either radiates outwards or converges to a specific point. Spiral paths do not contain well-defined boundaries compared to ripple patterns but do offer some energy savings due to having continuous motion with smaller changes to trajectory. These are commonly used for search missions requiring high-resolution coverage of a specific point.

As demonstrated in Fig. 2(b), these atomic patterns can be further combined to plan multi-phase composite missions.

3.3 UI For Language To Mission Mapping

Real-time visual feedback is essential for interactive verification of language to mission translation. To achieve this, we develop a 2D mission visualizer in W2W that places the human language commands into mission paths on a map for confirmation. Specifically, we generate a Leaflet map [66]-based UI on the given GPS coordinates; we further integrate options for subsequent interactions on the map, such as icons, zoom level, and map movement or corrections.

Fig. 3 shows a sample W2W-generated map with icons representing the corresponding command tokens. All individual tokens can be visualized on separate layers for further mission adaptations by the user. The corresponding map on the AUV interface is also shown in Fig. 3; it offers the waypoints with additional information. These waypoints are then loaded onto the AUV for deployment.

Refer to caption
Refer to caption
Figure 3: The W2W-generated mission map (top) and the corresponding map loaded on the AUV mission interface (bottom) are visualized; The W2W mission text used to generate this is: Start at 38.7969°38.7969\degree N, 75.1538°75.1538\degree W, Circle for a turn at a radius of 1010 m in a clockwise direction at an altitude of 11 m. Move south 3030 m and then Move south 1010 m. Move south for 100100 m and then Track left for 100100 m at a spacing of 1414 m. End at 38.7968°38.7968\degree N, 75.1535°75.1535\degree W.
Refer to caption
Refer to caption
Refer to caption
Figure 4: Three sample missions: input texts, generated waypoints, and the corresponding 2D mission maps are shown.

4 Word2Wave: Model Training

4.1 Model Selection and Data Preparation

Structured mission command generation from an input paragraph is effectively a text-to-text translation task. Thus, we consider SLM pipelines that are effective for text translation. While popular architectures such as GPT and BERT are mainly suited for text generation and classification [21, 67], Seq2Seq (Sequence to Sequence) models are more suited for translation tasks [68]. For this reason, we choose the T5 architecture, a Seq2Seq model which provides SOTA performance in targeted ‘text-to-text’ translation tasks [68].

In particular, we adopt a smaller variant of the original T5, named T5-Small as it is lightweight and computationally efficient [17, 37]. It processes the input text by encoding the sequence and subsequently decoding it into a concise and coherent summary. We leverage its pre-trained weights and fine-tune it on W2W translation tasks following the language rules and command structures that are presented in Sec. 3.1. The T5-Small version used for W2W, is the checkpoint with 6060 M parameters. It leverages the advantages of SLMs and provide an easy-to-train platform for generating mission waypoints from natural language.

Refer to caption
Figure 5: Outlines for dataset preparation, model training, and deployment processes of the W2W framework are shown.

Dataset Generation. LLM/SLM-based subsea mission programming for marine robotics is a relatively new area of research. There are no large-scale datasets available for generalized model training. Instead, we take a prompt engineering approach using OpenAI’s ChatGPT-4o [21], which helps produce various ways to program a particular mission. Different manners of phrasing the same command within the prompt ensure that various speech patterns are captured for comprehensive training. Particular attention is given so that outputs from the prompt follow the structure of a valid mission. A total of 11101110 different mission samples are prepared for supervised training. We verified each sample and their paired commands to ensure that they represent valid subsea missions for AUV deployment.

4.2 Training And Hyperparameter Tuning

We use a randomized 8080-55 percentile split for training and validation; the remaining 15%15\% samples are used for testing. During training, each line in the dataset is inserted into a tokenizer to extract embedded information. The training is conducted over 6060 epochs on a machine with 3232 GB of RAM and a single Nvidia RTX 30603060 Ti GPU with 88 GB of memory. The training is halted when the validation loss consistently remains below a threshold of 0.20.2.

Note that our training pipeline is not intended for general-purpose text-to-text translation, rather a targeted learning on a specific set of mission vocabulary. We make sure that the supervised learning strictly adhered to the language structure of the mission commands outlined in Sec. 3.1. A few sample examples are shown in Fig. 4. The end-to-end training, hyperparameter tuning, and deployment processes are shown in Fig. 5.

Sample outputs. We show three additional sample outputs for W2W (similar examples as in Fig. 4) below.

Sample mission #1:

Start at 29.786°29.786\degree N, 84.469°84.469\degree W. Move at a bearing of 180°180\degree for a distance of 200200 m at a speed of 11 m/s at an altitude of 88 m. Move South 180180 m and then Move South 120120 m. Track left for 450450 m at a spacing of 140140 m. Move North for a distance of 200200m at a speed of 11 m/s at an depth of 11 m, then Adjust to a depth of 0 meters. End at 38.7968°38.7968\degree N, 75.1535°75.1535\degree W

Output  #1: [S, 29.78629.786, 84.469-84.469], [Mv, 180180, 200200, 1.01.0, a, 8.08.0], [Mv, 180180, 180180, 1.01.0, n, 0], [Mv, 180180, 120120, 1.01.0, n, 0], [Tr, l, 450450, 140140, n, 0], [Mv, 0, 200200, 1.01.0, d, 11],[Az, d, 0],[E, 38.796838.7968, 75.1535-75.1535].

Sample mission #2:

Start at 20.535°20.535\degree N, and 165.714°165.714\degree W, Move at a bearing of 197°197\degree for a distance of 290290  m at a speed of 55 m/s. Track left for a distance of 795795 m spaced by 402402 m at a depth of 1717 m. End at 20.535°20.535\degree N, and 109.844°109.844\degree W

Output  #2: [S, 20.53520.535, 165.714-165.714], [Mv, 197197, 290290, 55, n, 0], [Tr, l, 795795, 402402, d, 1717], [E, 59.62959.629, 109.844109.844]

Sample mission #3:

Start at 38.806°38.806\degree N, 75.100°75.100\degree W, Move at a bearing of 45°45\degree for a distance of 100100 m at a speed of 11 m/s at a depth of 11 m. Spiral 55 turns with a radius of 5050 m counterclockwise. Adjust to a depth of 0 m. End at 38.7968°38.7968\degree N, 75.200°75.200\degree W

Output  #3: [S, 38.80638.806, 75.100-75.100], [Mv, 4545, 100100, 11, d, 1010], [Sp, 55, 5050, ccw, n, 0], [Az, d, 0], [E, 38.796838.7968, 75.200-75.200]

5 Experimental Analyses

5.1 Language Model Evaluation

Language models differ significantly in performance for certain tasks based on their underlying architectures, learning objectives, and application-specific design choices. As mentioned earlier, we adapt a sequence-to-sequence learning pipeline based on the T5-Small model [17]. For performance baseline and ablation experiments, we investigate two other SOTA models of the same genre: BART-Large [23] and MarianMT [24]. We compare their performance based on model accuracy, robustness, and computational efficiency.

Evaluation metrics and setup. To perform this analysis in an identical setup, we train and validate each model on the same dataset (see Sec. 4.2) until validation loss reaches a plateau. For accuracy, we use two widely used metrics: BLEU (bilingual evaluation understudy) [69] and METEOR (metric for evaluation of translation with explicit ordering) [70]. BLEU metric offers a language-independent understanding of how close a predicted sentence is to its ground truth. METEOR additionally considers the order of words during evaluation. It evaluates machine translation output based on the harmonic mean of unigram precision and recall, with recall weighted higher than precision.

In addition, we compare their inference rates on a set of 200200 test samples on the same device (RTX 30603060 TI GPU with 3232 GB memory). It is measured as the average time taken to generate a complete mission map given the input sample. Lastly, model sizes based on the number of parameters are apparent from the respective computational graphs.

Table 2: Performance comparison based on the number of parameters (in millions), BLEU and METEOR scores (in {0,1}), and inference speed (in miliseconds).
  \cellcolorgray!10Models \cellcolorgray!10Params (\downarrow) \cellcolorgray!10BLEU (\uparrow) \cellcolorgray!10METEOR (\uparrow) \cellcolorgray!10Speed (\downarrow)
  BART-Large [23] 406.291 M 0.268 0.405 5122.7 ms
MarianMT [24] 73.886 M 0.913 0.773 136.7 ms
Ours (T5-Small) 60.506 M 0.879 0.813 70.5 ms
 

Quantitative performance analyses of SOTA. Table 2 summarizes the quantitative performance comparison; it demonstrates that the MarianMT and T5-Small models offer more accurate and consistent scores when trained on a targeted dataset compared to BART-Large. We hypothesize that LLMs like BART-Large need more comprehensive datasets and are suited for general-purpose learning. On the other hand, T5-Small has marginally lower BLEU scores compared to MarianMT, while it offers better METEOR values at a significantly faster inference rate. T5-Small only has about 6060 M parameters, offering 2×2\times faster runtime while ensuring comparable accuracy and robustness as MarianMT. Further inspection reveals that MarianMT often randomizes the order of generated mission commands. T5-Small does not suffer from these issues, demonstrating a better balance between robustness and efficiency.

Refer to caption
Figure 6: Qualitative performance of our T5-Small mission generation engine for the 77 language commands in W2W.

Qualitative performance analyses. We evaluate the qualitative outputs of T5-Small for W2W language commands based on the number of inaccurate tokens generation across the whole test set of 200200 samples. As Fig. 6 shows, we categorize these into: missed tokens (failed to generate), erroneous tokens (incorrectly generated), and hallucinated (extraneously generated) tokens. Of all the error types, we found that the Adjust commands are hallucinated at a disproportionately greater rate. We hypothesize that it happens due to some bias learned by the model, causing it to associate changes in depth or altitude as an Adjust command. Besides, we observed relatively high missed token counts for Move and Circle commands. Track is another challenging command that suffers from high error rates for token generation. Nevertheless, 8989% tokens are accurately parsed from unseen examples, which we found to be enough for real-time mission programming by human participants.

Table 3: A SUS (system usability scale) [25] evaluation is conducted for 1010 usability questions, scaled from 1 (strongly disagree) to 5 (strongly agree). A total of 1515 user responses are compiled as mean (std. deviation) for our W2W interface and the NemoSens AUV interface (as baseline).
                                                            Q1: I think that I would like to use this system frequently.
Q2: I found the system unnecessarily complex.
Q3: I thought the system was easy to use.
Q4: I would need the support of a technical person to be able to use this system.
Q5: I found the various functions in this system were well integrated.
Q6: I thought there was too much inconsistency in this system.
Q7: Most people would learn to use this system very quickly.
Q8: I found the system very cumbersome to use.
Q9: I felt very confident using the system.
Q10: I needed to learn a lot of things before I could get going with this system.
  Q# \cellcolorgray!10Baseline \cellcolorgray!10W2W Interface Q# \cellcolorgray!10Baseline \cellcolorgray!10W2W Interface
  1 2.5 (1.4) 4.1 (0.8) 6 3.7 (0.9) 2.9 (1.0)
2 3.4 (0.9) 1.5 (0.8) 7 2.8 (1.0) 4.4 (1.2)
3 2.4 (0.9) 4.4 (0.8) 8 3.1 (1.2) 1.6 (0.9)
4 4.0 (1.5) 1.9 (0.7) 9 2.6 (1.0) 3.6 (1.2)
5 2.4 (0.8) 3.7 (1.0) 10 3.5 (0.9) 1.8 (1.0)
                                                     Metric       \cellcolorgray!10Baseline       \cellcolorgray!10W2W Interface
                                                    System usability       37.537.5 (16.9816.98)       76.2576.25 (19.4819.48)
Time taken (min:sec) 2727:0202 (0:3232) 0202:3737 (0:0303)
 

5.2 User Interaction Study

We assess the usability benefits of W2W compared to the NemoSens AUV interface, which we consider as the baseline. A total of 1515 individuals between the ages of 1818 to 3636 participated in our study; three of them were familiar with subsea mission programming and deployments, whereas the other 1212 people had no prior experience.

Evaluation procedure. Individuals were first introduced to both the W2W UI and Nemosens AUV interface for subsea mission programming. Then, they are asked to program three separate missions on these interfaces. As a quantitative measure, the total time taken to program each mission was recorded; they also completed a user survey to evaluate the degree of user satisfaction and ease of use between the two interfaces. This survey form was based on the system usability scale (SUS) [25]; the results are in Table 3.

User preference analyses. The participants rated W2W’s system usability at 76.2576.25, more than twice the SUS score of the baseline AUV programming interface. The participants generally expressed that the baseline interface is more complex, thus asked for assistance repeatedly. They reported several standard deviations higher scores for various features of W2W. They took less than 10%10\% time for programming missions on W2W, validating that it is more user-friendly and easier to use than the baseline.

5.3 Subsea Deployment Scenarios

We use a NemoSens AUV [19] for subsea mission deployments. It is a torpedo-shaped single-thruster AUV equipped with a DVL wayfinder, a down-facing HD camera, and a 450450 KHz side-scan sonar. As mentioned, the integrated software interface allows us to program various missions and generate waypoints for the AUV to execute in real-time. Once programmed, these waypoints are loaded to the AUV for deployment near the starting location.

Refer to caption
Refer to caption
Figure 7: The mission maps (left) and its sonar data (right) for a deployment at Delaware Bay (Golden Eagle shipwreck). The W2W mission command is: Start at 38.8670°38.8670\degree N, 75.1356°75.1356\degree W, Move south for 3030m, Adjust to a depth of 7.57.5m. Move south 3030 m and then Move east 120120 m. Track right for 100100 m with a spacing of 100100 m and then Move north for 100100 m. Track right for 120120 m with a spacing of 120120 m. Move south 2020 m and then Adjust to a depth of 2m. End at 38.867°38.867\degree N, 75.1642°75.1642\degree W..

In this general practice, W2W interface is intended as an HMI bridge, transferring the human language into the mission map; the rest of the experimental scenarios remain the same. The mission vocabulary and language rules adopted in W2W correspond to valid subsea missions programmed on an actual robot platform. We demonstrate this with several examples from our subsea deployments in the GoM (Gulf of Mexico) and the Delaware Bay, Atlantic Ocean. Real deployments are shown in Fig. 7 and earlier in Fig. 1.

As demonstrated in Fig. 7, targeted inspection or mapping missions can be programmed with only a few basic language commands and then validated on our 2D interface. In particular, the Track and Move commands are powerful tools that allow programming lawnmower and polygonal patterns fairly easily. Besides, Spiral and Circle commands are suited for more localized mission patterns. These intuitions are consistent with our user study evaluations as well.

Limitations and future work. Despite the accuracy and ease of use, W2W is a single-shot language summarizer, lacking human-machine dialogue capabilities. As discussed earlier in qualitative analyses (see Fig. 6), a major limitation is that W2W often generates hallucinated commands. Hence, the users must try again since the 2D mission maps show inaccurate patterns. This mainly happens when humans speak (instead of type) - as the standard speech-to-text translation packages often generate erroneous texts, which in turn generate wrong mission parameters. A sample failure case is illustrated in Fig. 8. W2W does not provide a method to edit any minor mistakes or specific parameters; hence users need to redo the program from the start.

As such, we are working on extending the mission programming engine to allow multi-shot generative features to enable memory and dialogues. This would provide the interactivity needed to make corrections to missions if needed. Our overarching goal to be able to use W2W to seek suggestions such as,

We want to map a shipwreck at 38.8670°38.8670\degree N, 75.1356°75.1356\degree W location. The wreck is about 200200 ft long, the altitude clearance is 5050 ft. Suggest a few mission patterns for mapping the shipwreck and its surroundings.

Users may engage in subsequent dialogue to find the best possible mission for their intended application, taking advantage of the W2W engine’s knowledge mined from comprehensive mission databases. We will integrate these embodied reasoning and dialogue-based mission programming features in future W2W versions.

Refer to caption
Refer to caption
Figure 8: A sample speech input from a human operator and the corresponding command list interpreted by Word2Wave are shown in the top two rows. Highlighted in red are words incorrectly parsed by the speech-to-tex translation model. Super-scripted texts in purple denote the hallucinated commands by the W2W. On the bottom row, we compare the respective mission maps generated by our W2W and the AUV interface.

6 Conclusions

This paper presents an SLM-driven mission programming framework named Word2Wave (W2W) for marine robotics. The use of natural language allows intuitive programming of subsea AUVs for remote mission deployments. We formulate novel language rules and intuitive command structures for efficient language-to-mission mapping in real-time. We develop a sequence-to-sequence learning pipeline based on T5-Small model, which demonstrates robust and more efficient performance compared to other SLM/LLM architectures. We also develop a mission map visualizer for users to validate the W2W-generated mission map before deployment. Through comprehensive quantitative, qualitative, and a user interaction study to validate the effectiveness of W2W through real subsea deployment scenarios. In future W2W versions, we will incorporate human-machine dialogue and spontaneous question-answering for embodied mission reasoning. We also intend to explore language-driven HMIs and adapt W2W capabilities for subsea telerobotics applications.

Acknowledgements

This work is supported in part by the National Science Foundation (NSF) grants #23304162330416 and #23261592326159. We are thankful to Dr. Arthur Trembanis, Dr. Herbert Tanner, and Dr. Kleio Baxevani at the University of Delaware for facilitating our field trials at the 2024 Autonomous Systems Bootcamp. We also acknowledge the system usability scale (SUS) study participants for helping us conduct the UI evaluation comprehensively.

References

  • [1] J. Wang, Z. Wu, Y. Li, H. Jiang, P. Shu, E. Shi, H. Hu, C. Ma, Y. Liu, X. Wang, Y. Yao, X. Liu, H. Zhao, Z. Liu, H. Dai, L. Zhao, B. Ge, X. Li, T. Liu, and S. Zhang, “Large Language Models for Robotics: Opportunities, Challenges, and Perspectives,” arXiv preprint arXiv:2401.04334, 2024.
  • [2] C. Zhang, J. Chen, J. Li, Y. Peng, and Z. Mao, “Large Language Models for Human-Robot Interaction: A Review,” Biomimetic Intelligence and Robotics, p. 100131, 2023.
  • [3] M. J. Islam, M. Ho, and J. Sattar, “Dynamic Reconfiguration of Mission Parameters in Underwater Human-Robot Collaboration,” in IEEE International Conference on Robotics and Automation (ICRA), pp. 1–8, 2018.
  • [4] N. Lucas Martinez, J.-F. Martínez-Ortega, P. Castillejo, and V. Beltran Martinez, “Survey of Mission Planning and Management Architectures for Underwater Cooperative Robotics Operations,” Applied Sciences, vol. 10, no. 3, p. 1086, 2020.
  • [5] J. Sattar and G. Dudek, “Underwater Human-robot Interaction via Biological Motion Identification,” in Robotics: Science and Systems (RSS), 2009.
  • [6] M. J. Islam, M. Ho, and J. Sattar, “Understanding Human Motion and Gestures for Underwater Human–Robot Collaboration,” Journal of Field Robotics (JFR), vol. 36, no. 5, pp. 851–873, 2019.
  • [7] A. Vasilijević, D. Nad, F. Mandić, N. Mišković, and Z. Vukić, “Coordinated Navigation of Surface and Underwater Marine Robotic Vehicles for Ocean Sampling and Environmental Monitoring,” IEEE/ASME transactions on mechatronics, vol. 22, no. 3, pp. 1174–1184, 2017.
  • [8] C. M. Gussen, C. Laot, F.-X. Socheleau, B. Zerr, T. Le Mézo, R. Bourdon, and C. Le Berre, “Optimization of Acoustic Communication Links for a Swarm of AUVs: The COMET and NEMOSENS Examples,” Applied Sciences, vol. 11, no. 17, p. 8200, 2021.
  • [9] J. Zhou, Y. Si, and Y. Chen, “A Review of Subsea AUV Technology,” Journal of Marine Science and Engineering, vol. 11, p. 1119, 2023.
  • [10] J. Lin, H. Gao, R. Xu, C. Wang, L. Guo, and S. Xu, “The Development of LLMs for Embodied Navigation,” arXiv preprint arXiv:2311.00530, 2023.
  • [11] S.-M. Park and Y.-G. Kim, “Visual Language Navigation: A Survey and Open Challenges,” Artificial Intelligence Review, vol. 56, no. 1, pp. 365–427, 2023.
  • [12] Z. Shao, Z. Yu, M. Wang, and J. Yu, “Prompting Large Language Models with Answer Heuristics for Knowledge-Based Visual Question Answering,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 14974–14983, 2023.
  • [13] T. Brooks, A. Holynski, and A. A. Efros, “Instructpix2pix: Learning to Follow Image Editing Instructions,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 18392–18402, 2023.
  • [14] W. Huang, F. Xia, T. Xiao, H. Chan, J. Liang, P. Florence, A. Zeng, J. Tompson, I. Mordatch, Y. Chebotar, P. Sermanet, N. Brown, T. Jackson, L. Luu, S. Levine, K. Hausman, and B. Ichter, “Inner Monologue: Embodied Reasoning Through Planning with Language Models,” arXiv preprint arXiv:2207.05608, 2022.
  • [15] R. Yang, M. Hou, J. Wang, and F. Zhang, “OceanChat: Piloting Autonomous Underwater Vehicles in Natural Language,” arXiv preprint arXiv:2309.16052, 2023.
  • [16] Z. Bi, N. Zhang, Y. Xue, Y. Ou, D. Ji, G. Zheng, and H. Chen, “OceanGPT: A Large Language Model for Ocean Science Tasks,” arXiv preprint arXiv:2310.02031, 2023.
  • [17] C. Raffel, N. Shazeer, A. Roberts, K. Lee, S. Narang, M. Matena, Y. Zhou, W. Li, and P. J. Liu, “Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer,” Journal of Machine Learning Research, vol. 21, no. 140, pp. 1–67, 2020.
  • [18] G. S. Andy, L. Mathieu, G. Bill, and L. Damian, “Subsea Robotics-AUV Pipeline Inspection Development Project,” in Offshore Technology Conference, p. D021S024R001, OTC, 2022.
  • [19] R. Inc., “NemoSens: An Open-Architecture, Cost-Effective, and Modular Micro-AUV.” https://rtsys.eu/nemosens-micro-auv, 2020.
  • [20] J. McMahon and E. Plaku, “Mission and Motion Planning for Autonomous Underwater Vehicles Operating in Spatially and Temporally Complex Environments,” IEEE Journal of Oceanic Engineering, vol. 41, no. 4, pp. 893–912, 2016.
  • [21] J. Achiam, S. Adler, S. Agarwal, L. Ahmad, I. Akkaya, F. L. Aleman, D. Almeida, J. Altenschmidt, S. Altman, S. Anadkat, et al., “Gpt-4 Technical Report,” arXiv preprint arXiv:2303.08774, 2023.
  • [22] A. Palnitkar, R. Kapu, X. Lin, C. Liu, N. Karapetyan, and Y. Aloimonos, “ChatSim: Underwater Simulation with Natural Language Prompting,” in OCEANS 2023-MTS/IEEE US Gulf Coast, pp. 1–7, IEEE, 2023.
  • [23] M. Lewis, “BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension,” arXiv preprint arXiv:1910.13461, 2019.
  • [24] M. Junczys-Dowmunt, R. Grundkiewicz, T. Dwojak, H. Hoang, K. Heafield, T. Neckermann, F. Seide, U. Germann, A. Fikri Aji, N. Bogoychev, A. F. T. Martins, and A. Birch, “Marian: Fast Neural Machine Translation in C++,” in Proceedings of ACL, System Demonstrations, pp. 116–121, 2018.
  • [25] J. Brooke, “SUS- A Quick and Dirty Usability Scale,” Usability Evaluation in Industry, vol. 189, no. 194, pp. 4–7, 1996.
  • [26] M. Chandarana, E. L. Meszaros, A. Trujillo, and B. D. Allen, “’Fly Like This’: Natural Language Interface for UAV Mission Planning,” in International Conference on Advances in Computer-Human Interactions, no. NF1676L-26108, 2017.
  • [27] S. García, P. Pelliccione, C. Menghi, T. Berger, and T. Bures, “High-Level Mission Specification for Multiple Robots,” in Proceedings of the 12th ACM SIGPLAN International Conference on Software Language Engineering, pp. 127–140, 2019.
  • [28] D. Shah, B. Eysenbach, G. Kahn, N. Rhinehart, and S. Levine, “Ving: Learning Open-World Navigation with Visual Goals,” in IEEE International Conference on Robotics and Automation (ICRA), pp. 13215–13222, 2021.
  • [29] G. Huang, “Visual-Inertial Navigation: A Concise Review,” in International Conference on Robotics and Automation (ICRA), pp. 9572–9582, IEEE, 2019.
  • [30] M. Chandarana, E. L. Meszaros, A. Trujillo, and B. D. Allen, “Challenges of Using Gestures in Multimodal HMI for Unmanned Mission Planning,” in Proceedings of the AHFE 2017 International Conference on Human Factors in Robots and Unmanned Systems, Springer, 2018.
  • [31] M. A. Shah, “A Language-Based Software Framework for Mission Planning in Autonomous Mobile Robots,” Master’s thesis, Pennsylvania State University, 2011.
  • [32] D. W. Mutschler, “Language Based Simulation, Flexibility, and Development Speed in the Joint Integrated Mission Model,” in Proceedings of the Winter Simulation Conference., p. 8, IEEE, 2005.
  • [33] D. C. Silva, P. H. Abreu, L. P. Reis, and E. Oliveira, “Development of a Flexible Language for Mission Description for Multi-Robot Missions,” Information Sciences, vol. 288, pp. 27–44, 2014.
  • [34] A. C. Trujillo, J. Puig-Navarro, S. B. Mehdi, and A. K. McQuarry, “Using Natural Language to Enable Mission Managers to Control Multiple Heterogeneous UAVs,” in Proceedings of the AHFE 2016 International Conference on Human Factors in Robots and Unmanned Systems, pp. 267–280, Springer, 2017.
  • [35] F. D. S. Cividanes, M. G. V. Ferreira, and F. de Novaes Kucinskis, “An Extended HTN Language for Onboard Planning and Acting Applied to a Goal-Based Autonomous Satellite,” IEEE Aerospace and Electronic Systems Magazine, vol. 36, no. 8, pp. 32–50, 2021.
  • [36] J. Wu, R. Antonova, A. Kan, M. Lepert, A. Zeng, S. Song, J. Bohg, S. Rusinkiewicz, and T. Funkhouser, “Tidybot: Personalized Robot Assistance with Large Language Models,” Autonomous Robots, vol. 47, no. 8, pp. 1087–1102, 2023.
  • [37] T. Kwon, N. Di Palo, and E. Johns, “Language Models as Zero-Shot Trajectory Generators,” IEEE Robotics and Automation Letters, 2024.
  • [38] D. Shah, B. Osiński, B. Ichter, and S. Levine, “LM-Nav: Robotic Navigation with Large Pre-trained Models of Language, Vision, and Action,” in Conference on Robot Learning, pp. 492–504, 2023.
  • [39] J. Gao, B. Sarkar, F. Xia, T. Xiao, J. Wu, B. Ichter, A. Majumdar, and D. Sadigh, “Physically Grounded Vision-Language Models for Robotic Manipulation,” in IEEE International Conference on Robotics and Automation (ICRA), pp. 12462–12469, 2024.
  • [40] Z. Mandi, S. Jain, and S. Song, “Roco: Dialectic Multi-robot Collaboration with Large Language Models,” in IEEE International Conference on Robotics and Automation (ICRA), pp. 286–299, 2024.
  • [41] P. J. B. Sánchez, M. Papaelias, and F. P. G. Márquez, “Autonomous Underwater Vehicles: Instrumentation and Measurements,” IEEE Instrumentation & Measurement Magazine, vol. 23, pp. 105–114, 2020.
  • [42] D. Wei, P. Hao, and H. Ma, “Architecture Design and Implementation of UUV Mission Planning,” in 2021 40th Chinese Control Conference (CCC), pp. 1869–1873, IEEE, 2021.
  • [43] I. A. McManus, A Multidisciplinary Approach to Highly Autonomous UAV Mission Planning and Piloting for Civilian Airspace. PhD thesis, Queensland University of Technology, 2005.
  • [44] P. Oliveira, A. Pascoal, V. Silva, and C. Silvestre, “Mission Control of the MARIUS Autonomous Underwater Vehicle: System Design, Implementation, and Sea trials,” International journal of systems science, vol. 29, no. 10, pp. 1065–1080, 1998.
  • [45] J. González-García, A. Gómez-Espinosa, E. Cuan-Urquizo, L. G. García-Valdovinos, T. Salgado-Jiménez, and J. A. Escobedo Cabello, “Autonomous Underwater Vehicles: Localization, Navigation, and Communication for Collaborative Missions,” Applied Sciences, vol. 10, no. 4, p. 1256, 2020.
  • [46] C. Whitt, J. Pearlman, B. Polagye, F. Caimi, F. Muller-Karger, A. Copping, H. Spence, S. Madhusudhana, W. Kirkwood, L. Grosjean, et al., “Future Vision for Autonomous Ocean Observations,” Frontiers in Marine Science, vol. 7, p. 697, 2020.
  • [47] S. S. Enan, M. Fulton, and J. Sattar, “Robotic Detection of a Human-comprehensible Gestural Language for Underwater Multi-human-robot Collaboration,” in 2022 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 3085–3092, IEEE, 2022.
  • [48] A. Abdullah, R. Chen, I. Rekleitis, and M. J. Islam, “Ego-to-Exo: Interfacing Third Person Visuals from Egocentric Views in Real-time for Improved ROV Teleoperation,” in In Review. ArXiv: 2407.00848, 2024.
  • [49] F. Xu, Q. Zhu, S. Li, Z. Song, and J. Du, “VR-Based Haptic Simulator for Subsea Robot Teleoperations,” in Computing in Civil Engineering, pp. 1024–1032, 2021.
  • [50] N. J. Hallin, B. Johnson, H. Egbo, M. O’Rourke, and D. Edwards, “Using Language-Centered Intelligence to Optimize Mine-Like Object Inspections for a Fleet of Autonomous Underwater Vehicles,” UUST’09, 2009.
  • [51] M. J. Islam, J. Mo, and J. Sattar, “Robot-to-robot Relative Pose Estimation using Humans as Markers,” Autonomous Robots, vol. 45, no. 4, pp. 579–593, 2021.
  • [52] J. Wu, B. Yu, and M. J. Islam, “3d reconstruction of underwater scenes using nonlinear domain projection,” in 2023 IEEE Conference on Artificial Intelligence (CAI), pp. 359–361, IEEE, 2023. Best Paper Award.
  • [53] R. Capocci, G. Dooly, E. Omerdić, J. Coleman, T. Newe, and D. Toal, “Inspection-Class Remotely Operated Vehicles—A Review,” Journal of Marine Science and Engineering, vol. 5, no. 1, p. 13, 2017.
  • [54] VideoRay, “OctoView: A Cutting Edge Mixed Reality Interface for Pilots and Teams.” https://videoray.com/labs/, 2024.
  • [55] M. J. Islam, “Eye On the Back: Augmented Visuals for Improved ROV Teleoperation in Deep Water Surveillance and Inspection,” in SPIE Defense and Commercial Sensing, (Maryland, USA), SPIE, 2024.
  • [56] F. Xu, T. Nguyen, and J. Du, “Augmented Reality for Maintenance Tasks with ChatGPT for Automated Text-to-Action,” Journal of Construction Engineering and Management, vol. 150, no. 4, 2024.
  • [57] M. Laranjeira, A. Arnaubec, L. Brignone, C. Dune, and J. Opderbecke, “3D Perception and Augmented Reality Developments in Underwater Robotics for Ocean Sciences,” Current Robotics Reports, vol. 1, pp. 123–130, 2020.
  • [58] F. Gao, “Mission Planning and Replanning for ROVs,” Master’s thesis, NTNU, 2020.
  • [59] J. E. Manley, S. Halpin, N. Radford, and M. Ondler, “Aquanaut: A New Tool for Subsea Inspection and Intervention,” in OCEANS 2018 MTS/IEEE Charleston, pp. 1–4, IEEE, 2018.
  • [60] P. Parekh, C. McGuire, and J. Imyak, “Underwater robotics semantic parser assistant,” arXiv preprint arXiv:2301.12134, 2023.
  • [61] C. R. Samuelson and J. G. Mangelson, “A Guided Gaussian-Dirichlet Random Field for Scientist-in-the-Loop Inference in Underwater Robotics,” in 2024 IEEE International Conference on Robotics and Automation (ICRA), pp. 9448–9454, IEEE, 2024.
  • [62] I. C. Rankin, S. McCammon, and G. A. Hollinger, “Robotic Information Gathering Using Semantic Language Instructions,” in 2021 IEEE International Conference on Robotics and Automation (ICRA), pp. 4882–4888, IEEE, 2021.
  • [63] E. Potokar, S. Ashford, M. Kaess, and J. G. Mangelson, “HoloOcean: An Underwater Robotics Simulator,” in International Conference on Robotics and Automation (ICRA), pp. 3040–3046, IEEE, 2022.
  • [64] M. M. M. Manhães, S. A. Scherer, M. Voss, L. R. Douat, and T. Rauschenbach, “UUV Simulator: A Gazebo-Based Package for Underwater Intervention and Multi-Robot Simulation,” in Oceans MTS/IEEE Monterey, pp. 1–8, 2016.
  • [65] M. Prats, J. Perez, J. J. Fernández, and P. J. Sanz, “An Open Source Tool for Simulation and Supervision of Underwater Intervention Missions,” in IEEE/RSJ International Conference on Intelligent Robots and Systems, pp. 2577–2582, 2012.
  • [66] V. Agafonkin, “Leaflet.” https://leafletjs.com/, 2010.
  • [67] J. Devlin, “BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding,” arXiv preprint arXiv:1810.04805, 2018.
  • [68] A. Choudhary, M. Alugubelly, and R. Bhargava, “A Comparative Study on Transformer-based News Summarization,” in International Conference on Developments in eSystems Engineering (DeSE), pp. 256–261, 2023.
  • [69] K. Papineni, S. Roukos, T. Ward, and W.-J. Zhu, “BLEU: A Method for Automatic Evaluation of Machine Translation,” in Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics, pp. 311–318, 2002.
  • [70] S. Banerjee and A. Lavie, “METEOR: An Automatic Metric for MT Evaluation with Improved Correlation with Human Judgments,” in Proceedings of the ACL Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization, pp. 65–72, 2005.