Decoupled Audio-Visual Dataset Distillation

Anya Sharma Ben Carter Chen Li
School of Electrical Engineering and Computer Science, University of Applied Sciences, Metropolis, Country

Abstract

This paper introduces a novel decoupled approach for audio-visual dataset distillation, aiming to address the inherent complexities and interdependencies in multimodal data. We propose a framework that independently distills audio and visual knowledge before a refined fusion step, leading to more efficient and effective student models. Experimental results demonstrate that our decoupled method significantly outperforms conventional coupled distillation techniques across various benchmarks, achieving higher performance with reduced computational overhead. This research provides a promising direction for scaling knowledge distillation to large and intricate multimodal datasets.

Keywords

Audio-Visual Learning, Dataset Distillation, Knowledge Distillation, Multimodal AI, Decoupled Architectures


1. Introduction

The proliferation of large-scale audio-visual datasets presents significant challenges for efficient model training and deployment, necessitating effective knowledge compression techniques like dataset distillation. Traditional distillation methods often struggle with the inherent coupling and complex feature interactions in multimodal data, leading to suboptimal performance or increased computational costs. This work proposes a novel decoupled strategy to enhance the distillation process for audio-visual datasets by handling modalities independently before integration. Models used in this article include a Teacher-Student Network, a Modality-Specific Feature Encoder (for both audio and visual streams), a Decoupling Knowledge Transfer Module, and a Multimodal Fusion Head.

2. Related Work

Prior research in dataset distillation primarily focuses on unimodal data, employing techniques like matching gradients or feature distributions to synthesize small proxy datasets. While multimodal learning has seen advancements in fusion strategies and cross-modal understanding, their integration with dataset distillation remains a challenge due to the complexities of inter-modal dependencies. Existing audio-visual distillation attempts often treat modalities as a single input, overlooking the potential benefits of independent processing and specific knowledge transfer pathways. Our work differentiates by explicitly decoupling the distillation process for enhanced efficiency and performance.

3. Methodology

Our proposed methodology involves a two-stage decoupled distillation process for audio-visual datasets. Initially, separate teacher models distill modality-specific knowledge into distinct audio and visual student networks, creating compact synthetic datasets or knowledge representations for each stream. Subsequently, a multimodal fusion module is trained on these independently distilled components, learning optimal integration strategies without the direct interference of highly coupled raw data. This approach minimizes cross-modal noise during initial distillation and allows for more focused learning within each modality, improving overall knowledge transfer efficiency.

4. Experimental Results

Experiments conducted on established audio-visual classification benchmarks demonstrate the superior performance of our decoupled distillation approach compared to traditional coupled methods. The decoupled strategy consistently yielded higher classification accuracies and required significantly smaller distilled datasets while maintaining competitive training times. For instance, on the AudioSet dataset, our method achieved a 3% accuracy improvement over the best coupled baseline with a 5x smaller distilled dataset.

Table I. Performance Comparison of Distillation Methods

Method Accuracy (%) Distillation Ratio Training Time (hrs)
Baseline Coupled Distillation 78.2 1:100 15.5
Feature-Matching Coupled Distillation 79.5 1:80 16.2
Proposed Decoupled Distillation 82.5 1:500 14.0
The table above clearly illustrates that the proposed decoupled distillation method not only achieves a higher accuracy but also significantly reduces the required distillation ratio, indicating greater efficiency in knowledge transfer. Furthermore, it maintains a competitive training time, proving its practical applicability.

5. Discussion

The improved performance achieved by the decoupled audio-visual dataset distillation highlights the benefits of specialized knowledge transfer for each modality prior to integration. This strategy effectively mitigates the challenges of high-dimensional, inter-dependent multimodal data, allowing student models to learn more robust and generalized representations. The implications of this work extend to various resource-constrained environments where efficient model deployment is critical, potentially enabling the use of complex audio-visual models on edge devices. Future work could explore adaptive decoupling strategies and extend this framework to other multimodal learning tasks.