This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Integrating Audio Narrations to Strengthen Domain Generalization in Multimodal First-Person Action Recognition

Abstract

First-person activity recognition is rapidly growing due to the widespread use of wearable cameras but faces challenges from domain shifts across different environments, such as varying objects or background scenes. We propose a multimodal framework that improves domain generalization by integrating motion, audio, and appearance features. Key contributions include analyzing the resilience of audio and motion features to domain shifts, using audio narrations for enhanced audio-text alignment, and applying consistency ratings between audio and visual narrations to optimize the impact of audio in recognition during training. Our approach achieves state-of-the-art performance on the ARGO1M dataset, effectively generalizing across unseen scenarios and locations.

Index Terms—  Action recognition, multimodal domain generalization, audio descriptions, multimodal fusion

1 Introduction

With the growing prevalence of wearable technology and first-person cameras, first-person activity recognition has emerged as a crucial area of research [1, 2, 3, 4, 5]. This field is vital for real-world egocentric vision applications, from human-robot interaction to personalized assistance. However, a significant challenge in developing robust action recognition models is the issue of domain shifts—variations in environmental contexts, objects, and activities that can drastically affect the performance of these models. Some works [6, 7, 8, 9, 10] rely solely on visual features to generalize across different domains. In this paper, we explore the potential of multimodality, specifically the integration of motion and audio with appearance, to enhance domain generalization in first-person action recognition tasks.

Refer to caption
Fig. 1: Illustration of motion and audio resilience to domain shifts compared to appearance. While the motion of ‘cutting’ (first row) and audio of ‘pouring’ (second row) remain similar across different scenario-location domains, the appearance varies significantly with different objects and backgrounds.

Our intuition is that domain shifts arise heavily from variations in the spatial semantics of videos, such as changes in the type of objects or the background, which we refer to as the appearance modality. These variations can lead to a drop in model performance when applied to new, unseen data. However, the motion and audio modalities capture temporal dynamics that remain more consistent across different domains.

Refer to caption
Fig. 2: The proposed framework extracts appearance apiap_{i}, motion mim_{i}, and audio aia_{i} embeddings using trained encoders ff. Visual-text and audio-text alignments are performed independently to enhance the robustness of action representations. Consistency rating rir_{i}, calculated offline using a LLM [11], is then multiplied by audio embedding, optimizing the influence of audio in multimodal prediction. Note that narrations and consistency rate are only utilized during training to improve representation learning. During inference, embeddings are directly fused before prediction.

For example, while objects being ‘cut’ in different domains, such as bamboo stick with knife (in building in India scenario) or cable wire with pliers (in the mechanic in Pennsylvania scenario) are visually different, the motion involved in cutting, characterized by the repetitive back-and-forth movement, remains consistent (Fig. 1 top). Similarly, ‘mixing’ motions are consistent whether stirring ingredients or mixing cement. Audio also provides strong temporal cues that are less affected by visual changes. The sound of ‘pouring’, whether paint into a tray or water from glass, remains consistent (Fig. 1 bottom). Similarly, the sound of ‘rolling’, whether dough in a kitchen or clay in a studio, stays the same. Despite variations in visual context, the consistent motion patterns and audio cues are more reliable indicators of the action and are more robust to domain shifts. Our experimental analysis validate our intuition by showcasing lower performance drop of motion (25.8%) and audio (32.7%) comparing to appearance (54.8%) when applied to unseen domains.

In addition to leveraging motion and audio, our research also incorporates the text modality through narrations. Aligning visual features with corresponding textual descriptions (in this case, narrations of actions) has been shown to enhance feature representation [12, 10, 13]. However, existing narrations are primarily based on visual content (in [10]’s dataset, many videos do not contain audio, yet all contain narrations). Thus aligning audio features with them can introduce noise due to inconsistencies, as the actions depicted in the video may not always produce corresponding audio cues. To address this, we employ an audio captioner [14] to generate audio-specific narrations, enabling us to align appearance and motion with visual-based narrations, and audio with its respective narrations. Additionally, we calculate consistency ratings between visual and audio narrations using an LLM [11]. This consistency indicates how closely the audio and video express the same action. We apply attention weighing (Fig. 2) where lower consistency reduces the influence of audio in prediction. These approaches improve representation learning by reducing the noise and enhancing generalization.

Prior work in multimodal domain generalization for action recognition includes RNA-Net [15], which balances audio and video feature norms, and SimMMDG [16], which employs contrastive learning and cross-modal translation to separate modality-specific and shared features. In contrast, our approach provides experimental analysis that highlights the resilience of audio and motion features to domain shifts, emphasizing their pivotal role in achieving more robust generalization compared to prior methods. CIR [10] introduces ARGO1M for testing action recognition generalization across unseen scenarios and locations, using cross-instance reconstruction with text guidance, aligning visual narrations with appearance. Differently, we align audio with audio-based narrations, improving the robustness of representations. While [17] uses an LLM to generate audio-centric narrations from visual ones, we obtain audio narrations from audio with an audio captioner [14] and calculate consistency rate to determine how consistent audio is with video.

To summarize, our contributions are: (1) a multimodal framework that integrates motion, audio, and appearance features to advance domain generalization in first-person activity recognition, achieving state-of-the-art results on the ARGO1M dataset; (2) a thorough analysis highlighting the resilience of motion and audio features to domain shifts, emphasizing their pivotal role in domain generalization; (3) the alignment of audio narrations with audio features to strengthen the robustness of action representations; and (4) the use of consistency ratings between audio and visual narrations to optimize the influence of audio in prediction.

2 Method

2.1 Proposed Multimodal Setting

Each training sample consists of a video clip paired with corresponding audio, visual narration and audio narration. The former is provided in the dataset we use [10], while we use Pengi [14] to generate the latter. While we adopt the approach from [10] of using separate frozen encoders for each modality to extract base features, our method differs by incorporating the audio modality and utilizing distinct encoders for each modality, instead of a single encoder for fused appearance and motion features. Specifically, we train (from scratch) separate encoders fappearancef_{appearance}, fmotionf_{motion}, and faudiof_{audio} to derive domain-generalizable features apap, mm, and aa for appearance, motion, and audio. Additionally, we train a separate encoder ftextf_{text} extracts the features of visual narration tt and audio narration t^\hat{t}. See Fig. 2. The action prediction y^\hat{y} is generated by the classifier hh based on the fused multimodal embedding. The cross-entropy loss c\mathcal{L}_{c} is computed between the true and predicted action labels, yy and y^\hat{y}.

2.2 Consistency Ratings to Enhance Fusion

Before training, we calculate consistency ratings for each video sample using a LLM [11], assessing the degree to which the audio and visual narrations correspond semantically. This information is used to control fusion. Computing consistency between audio and visual content through text captures abstract concepts that raw features might miss. While features emphasize low-level details, text-based evaluations ensure that the audio and visual content correspond at a deeper, conceptual level.

We use the following prompt obtain consistency ratings:

Rate the consistency between two narrations from the same video out of 100. The first narration describes the visual aspect, and the second describes the audio. Consider how well the audio narration overlaps with and complements the visual narration. Output only the percentage score.

The consistency ratings rr are then used as weights to modulate the contribution of the audio embeddings aa before the concatenation of appearance, motion, and audio features, as shown in Fig. 2. Specifically, only during training, the audio embeddings are scaled by the consistency rating rr, such that a=ara=a*r. This consistency-weighted audio approach ensures that audio information with strong semantic alignment to the visual content exerts a greater influence on the final prediction. This approach enhances the quality of audio representations during training, minimizing the impact of noisy or irrelevant audio cues and improving the overall robustness.

2.3 Text Guided Alignment

Aligning visual features with narrations enriches the model with domain-invariant, human-like understanding, enhancing its ability to generalize across domains [13]. As discussed previously, we generate audio narrations that are specifically tied to the audio content. The LLM in audio captioner Pengi [14] mimics human-like understanding, providing audio-specific, semantically rich descriptions. Since Pengi uses a separate encoder for audio feature extraction, its generated audio narrations are initially not aligned with our model’s audio features aa. Thus, in our approach, we align audio features aa with these audio narration features t^\hat{t}, while aligning appearance features apap and motion features mm with visual narration features tt using contrastive learning.

Given a batch of samples ={(api,mi,ai,ti,t^i)}i=1B\mathcal{B}=\{(ap_{i},m_{i},a_{i},t_{i},\hat{t}_{i})\}^{B}_{i=1}, we frame the alignment as noise contrastive estimation [18]. Specifically, apt\mathcal{L}_{{ap}\rightarrow{t}} treats the appearance apiap_{i} as the anchor, with other narrative texts serving as negatives, and minimizes:

apt=1||iBlogexp(s(api,ti)/τ)jBexp(s(apj,tj)/τ)\mathcal{L}_{{ap}\rightarrow{t}}=-\dfrac{1}{|\mathcal{B}|}\sum^{B}_{i}log\dfrac{exp(s(ap_{i},t_{i})/\tau)}{\sum^{B}_{j}exp(s(ap_{j},t_{j})/\tau)} (1)

where s(,)s(\cdot,\cdot) is the cosine similarity and τ\tau is a learnable temperature. In a similar manner, tap\mathcal{L}_{{t}\rightarrow{ap}} uses the text tit_{i} as the anchor, with other appearance features as negatives. These losses are then combined to create the appearance-text alignment loss ap,t=tap+apt\mathcal{L}_{ap,t}=\mathcal{L}_{{t}\rightarrow{ap}}+\mathcal{L}_{{ap}\rightarrow{t}}. Likewise, motion-text alignment m,t\mathcal{L}_{m,t} and audio-text alignment a,t^\mathcal{L}_{a,\hat{t}} are computed, and all these alignment losses are aggregated into a total alignment loss align=ap,t+m,t+a,t^\mathcal{L}_{align}=\mathcal{L}_{ap,t}+\mathcal{L}_{m,t}+\mathcal{L}_{a,\hat{t}}.

To form the overall training objective, we combine the alignment loss with the cross-entropy classification loss:

=c+λalign\mathcal{L}=\mathcal{L}_{c}+\lambda\mathcal{L}_{align} (2)

where λ=0.1\lambda=0.1 is used to weight the alignment loss.

3 Experiments

Modality Me Co Ar Sh Cl Mean
Setting SAU JPN ITA IND US-MN
Audio* 27.6 42.9 30.4 29.3 28.7 31.8
Audio 25.8 23.1 23.0 24.9 24.5 24.3
(-6.9%) (-85.7%) (-32.1%) (-21.6%) (-17.1%) (-32.7%)
Motion* 27.0 29.3 32.0 29.6 26.2 28.9
Motion 25.9 19.1 21.9 26.8 22.8 23.3
(-4.2%) (-53.4%) (-46.1%) (-10.4%) (-14.9%) (-25.8%)
Appearance* 34.9 56.4 51.2 43.0 39.2 44.9
Appearance 31.6 26.6 29.3 31.1 28.7 29.5
(-10.4%) (-112.0%) (-74.7%) (-38.2%) (-38.2%) (-54.8%)
Multimodal* 36.5 59.0 52.5 44.9 40.4 46.7
Multimodal 34.7 30.7 31.8 34.2 32.4 32.7
(-5.1%) (-92.1%) (-65.1%) (-31.2%) (-24.7%) (-42.8%)
Table 1: Percentages show performance drop when the training set excludes samples from the test domain, vs when it includes them (*). Audio and motion exhibit less performance drop compared to appearance, highlighting their resilience to shifts. The baseline method (cross-entropy loss) is used.
Method Me Co Ar Sh Cl Pl Sp Ga Bu Kn Mean
SAU JPN ITA IND US-MN US-IN COL US-PNA US-PNA IND
Baseline 34.7 30.7 31.8 34.2 32.4 36.1 33.8 34.6 33.9 28.9 33.1
CIR [10] 35.6 32.4 32.4 35.4 33.1 38.2 34.7 36.1 35.4 30.4 34.3
Ours 36.0 32.2 32.7 36.0 33.5 38.3 35.1 36.5 35.8 30.6 34.7
Table 2: A comparison with the state of the art on ARGO1M using appearance, motion, and audio.
Method Modality Me Co Ar Sh Cl Mean
Setting SAU JPN ITA IND US-MN
Baseline Ap 31.6 26.6 29.3 31.1 28.7 29.5
Baseline Ap-Mo 32.7 27.5 30.0 31.3 29.5 30.2
Baseline Ap, Mo 33.2 27.8 30.8 31.7 29.7 30.6
Baseline Ap, Mo, Au 34.7 30.7 31.8 34.2 32.4 32.7
Table 3: Impact of combining different modalities—Appearance, Motion, and Audio—across domains.
Method Me Co Ar Sh Cl Mean
SAU JPN ITA IND US-MN
Baseline (BB) 34.7 30.7 31.8 34.2 32.4 32.7
B+B+ weighted aa by rr 34.9 30.9 31.9 34.4 32.4 32.9
B+vt+mt+atB+\mathcal{L}_{vt}+\mathcal{L}_{mt}+\mathcal{L}_{at} 35.4 31.6 32.3 35.2 32.9 33.5
B+vt+mt+at^B+\mathcal{L}_{vt}+\mathcal{L}_{mt}+\mathcal{L}_{a\hat{t}} 35.8 31.9 32.6 35.6 33.3 33.8
Ours 36.0 32.2 32.7 36.0 33.5 34.1
Table 4: Ablation showing the impact of alignment losses and consistency-weighted audio. All methods use (Ap, Mo, Au).

Implementation details. We leverage the frozen pretrained SlowFast model [19], utilizing its slow pathway for capturing appearance features and its fast pathway for extracting motion features. For audio features, we employ BEATs [20], while CLIP-ViT-B-32 [12] is used to extract text features. The trained encoders ff each consist of two fully connected layers, each followed by ReLU activation and Batch Normalization [21]. We use batch size 128 for 50 epochs with the Adam optimizer [22]. The learning rate is initially set to 2e-4, with decay factor 10 applied at epochs 30 and 40.

Dataset. ARGO1M [10] is an egocentric video dataset curated from Ego4D, specifically designed to analyze scenario and location-based domain shifts. The dataset includes 10 distinct train-test splits, ensuring that the domains of the samples—both scenario and location—do not overlap between the training and test sets. The test set scenarios and geographic locations include Gardening in Pennsylvania (Ga, US-PNA), Cleaning in Minnesota (Cl, US-MN), Knitting in India (Kn, IND), Shopping in India (Sh, IND), Building in Pennsylvania (Bu, US-PNA), Mechanic in Saudi Arabia (Me, SAU), Sport in Colombia (Sp, COL), Cooking in Japan (Co, JPN), Arts and Crafts in Italy (Ar, ITA), and Playing in Indiana (Pl, US-IN). ARGO1M contains 1,050,371 video clips. However, due to the absence of audio in some clips, we utilized approximately 60% of the total dataset where audio is available. We report top-1 accuracy for each test split, along with the mean accuracy across the splits.

3.1 Audio and Motion Resilience to Domain Shifts

In Table 1, the mean percentage drops in parentheses indicate the extent of domain shifts across various domains, comparing modality settings where the training set either contains samples that share location and/or scenario with the test domain, or share neither. The mean performance drop for motion features is 25.8%, while audio exhibits a drop of 32.7%. In contrast, appearance shows a significantly higher shift (54.8%). This aligns with our hypothesis that temporal dynamics—such as consistent patterns of movement and continuity of sound—remain more stable across different environments and scenarios. In contrast, spatial semantics represented by appearance features, are more variable due to differences in objects, backgrounds, and other visual elements that can vary significantly from one domain to another. Moreover, the multimodal approach, which integrates audio, motion, and appearance, demonstrates a reduced shift with a mean drop of 42.8% compared to appearance alone (54.8%). Our analysis underscores the critical role of incorporating audio and motion features for robust domain generalization.

3.2 Comparison with State of the Art

Table 2 compares our proposed method, the state-of-the-art CIR [10] approach, and a baseline model on the ARGO1M dataset. The baseline is trained solely with cross-entropy loss, whereas CIR enhances its performance through cross-instance reconstruction guided by text. Notably, although the original CIR results did not include audio, we have extended their approach by incorporating audio features and multiple trained encoders ff for each modality to ensure a fair comparison with our method. Our method consistently outperforms CIR across all domain splits, except for Co-JPN, achieving up to a 1.7% improvement and a 1.2% higher average performance overall. Additionally, our method outperforms the baseline by an average of 4.8%.

3.3 Ablations

Table 3 illustrates the impact of different modality settings on performance. In the ‘Ap-Mo’ setting used by CIR [10], appearance and motion features are fused early and processed through a single encoder. We propose an alternative approach in the ‘Ap, Mo’ setting, where separate encoders (fappearancef_{appearance}, fmotionf_{motion}) are employed to learn distinct domain-generalizable features, particularly because motion exhibits greater resistance to domain shifts. As a result, our ‘Ap, Mo’ outperforms ‘Ap-Mo’. Furthermore, the integration of audio with separate encoders in ‘Ap, Mo, Au’ yields the best overall performance.

Table  4 evaluates the effectiveness of each component of our approach. Starting with the baseline (trained using only cross-entropy), we observe that weighting the audio embeddings aa by the consistency ratings rr enhances the model’s performance, owing to the more reliable audio representations. Aligning all modalities with narration (‘B+vt+mt+atB+\mathcal{L}_{vt}+\mathcal{L}_{mt}+\mathcal{L}_{at}‘) further improves performance. Notably, when we align audio features with their corresponding audio narrations t^\hat{t} using at^\mathcal{L}_{a\hat{t}} the model outperforms the version with at\mathcal{L}_{at} where audio features are aligned with the original visual-based narrations tt. Finally, the combination of alignment losses and the consistency-weighted audio approach in our final method ‘Ours‘ achieves the best performance across all domains.

Conclusion. We proposed a novel multimodal framework where motion, audio, and appearance improve domain generalization in first-person action recognition. We demonstrated that audio and motion features exhibit greater resilience to domain shifts than appearance features. By aligning audio features with audio-specific narrations and applying consistency-weighted audio during training, our method enhanced the robustness of action representations.

References

  • [1] Evangelos Kazakos, Arsha Nagrani, Andrew Zisserman, and Dima Damen, “Epic-fusion: Audio-visual temporal binding for egocentric action recognition,” in ICCV, 2019, pp. 5492–5501.
  • [2] Xinyu Gong, Sreyas Mohan, Naina Dhingra, Jean-Charles Bazin, Yilei Li, Zhangyang Wang, and Rakesh Ranjan, “Mmg-ego4d: Multimodal generalization in egocentric action recognition,” in CVPR, 2023, pp. 6481–6491.
  • [3] Yue Zhao, Ishan Misra, Philipp Krähenbühl, and Rohit Girdhar, “Learning video representations from large language models,” in CVPR, 2023, pp. 6586–6597.
  • [4] Xiaohan Wang, Linchao Zhu, Heng Wang, and Yi Yang, “Interactive prototype learning for egocentric action recognition,” in ICCV, 2021, pp. 8168–8177.
  • [5] Chiara Plizzari, Mirco Planamente, Gabriele Goletto, Marco Cannici, Emanuele Gusso, Matteo Matteucci, and Barbara Caputo, “E2 (go) motion: Motion augmented event stream for egocentric action recognition,” in CVPR, 2022, pp. 19935–19947.
  • [6] Da Li, Yongxin Yang, Yi-Zhe Song, and Timothy M Hospedales, “Deeper, broader and artier domain generalization,” in ICCV, 2017, pp. 5542–5550.
  • [7] Krikamol Muandet, David Balduzzi, and Bernhard Schölkopf, “Domain generalization via invariant feature representation,” in ICML, 2013, pp. 10–18.
  • [8] Vihari Piratla, Praneeth Netrapalli, and Sunita Sarawagi, “Efficient domain generalization via common-specific low-rank decomposition,” in ICML, 2020, pp. 7728–7738.
  • [9] Hongyi Zhang, Moustapha Cisse, Yann N Dauphin, and David Lopez-Paz, “mixup: Beyond empirical risk minimization,” arXiv preprint arXiv:1710.09412, 2017.
  • [10] Chiara Plizzari, Toby Perrett, Barbara Caputo, and Dima Damen, “What can a cook in italy teach a mechanic in india? action recognition generalisation over scenarios and locations,” in ICCV, 2023, pp. 13656–13666.
  • [11] Hugo Touvron, Thibaut Lavril, Gautier Izacard, Xavier Martinet, Marie-Anne Lachaux, Timothée Lacroix, Baptiste Rozière, Naman Goyal, Eric Hambro, Faisal Azhar, et al., “Llama: Open and efficient foundation language models,” arXiv preprint arXiv:2302.13971, 2023.
  • [12] Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al., “Learning transferable visual models from natural language supervision,” in ICML, 2021, pp. 8748–8763.
  • [13] Seonwoo Min, Nokyung Park, Siwon Kim, Seunghyun Park, and Jinkyu Kim, “Grounding visual representations with texts for domain generalization,” in ECCV, 2022, pp. 37–53.
  • [14] Soham Deshmukh, Benjamin Elizalde, Rita Singh, and Huaming Wang, “Pengi: An audio language model for audio tasks,” NeurIPS, vol. 36, pp. 18090–18108, 2023.
  • [15] Mirco Planamente, Chiara Plizzari, Emanuele Alberti, and Barbara Caputo, “Domain generalization through audio-visual relative norm alignment in first person action recognition,” in WACV, 2022, pp. 1807–1818.
  • [16] Hao Dong, Ismail Nejjar, Han Sun, Eleni Chatzi, and Olga Fink, “Simmmdg: A simple and effective framework for multi-modal domain generalization,” NeurIPS, vol. 36, 2024.
  • [17] Andreea-Maria Oncescu, João F Henriques, Andrew Zisserman, Samuel Albanie, and A Sophia Koepke, “A sound approach: Using large language models to generate audio descriptions for egocentric text-audio retrieval,” in ICASSP, 2024, pp. 7300–7304.
  • [18] Aaron van den Oord, Yazhe Li, and Oriol Vinyals, “Representation learning with contrastive predictive coding,” arXiv preprint arXiv:1807.03748, 2018.
  • [19] Christoph Feichtenhofer, Haoqi Fan, Jitendra Malik, and Kaiming He, “Slowfast networks for video recognition,” in ICCV, 2019, pp. 6202–6211.
  • [20] Sanyuan Chen, Yu Wu, Chengyi Wang, Shujie Liu, Daniel Tompkins, Zhuo Chen, and Furu Wei, “Beats: Audio pre-training with acoustic tokenizers,” arXiv preprint arXiv:2212.09058, 2022.
  • [21] Sergey Ioffe, “Batch normalization: Accelerating deep network training by reducing internal covariate shift,” arXiv preprint arXiv:1502.03167, 2015.
  • [22] Diederik P Kingma, “Adam: A method for stochastic optimization,” arXiv preprint arXiv:1412.6980, 2014.