← Back to Timeline

Separating the "Chirp" from the "Chat": Self-supervised Visual Grounding of Sound and Language

Foundational AI

Authors

Mark Hamilton, Andrew Zisserman, John R. Hershey, William T. Freeman

Abstract

We present DenseAV, a novel dual encoder grounding architecture that learns high-resolution, semantically meaningful, and audio-visually aligned features solely through watching videos. We show that DenseAV can discover the ``meaning'' of words and the ``location'' of sounds without explicit localization supervision. Furthermore, it automatically discovers and distinguishes between these two types of associations without supervision. We show that DenseAV's localization abilities arise from a new multi-head feature aggregation operator that directly compares dense image and audio representations for contrastive learning. In contrast, many other systems that learn ``global'' audio and video representations cannot localize words and sound. Finally, we contribute two new datasets to improve the evaluation of AV representations through speech and sound prompted semantic segmentation. On these and other datasets we show DenseAV dramatically outperforms the prior art on speech and sound prompted semantic segmentation. DenseAV outperforms the previous state-of-the-art, ImageBind, on cross-modal retrieval using fewer than half of the parameters. Project Page: \href{https://aka.ms/denseav}{https://aka.ms/denseav}

Concepts

audio-visual grounding self-supervised learning contrastive learning dense cross-modal alignment representation learning disentangled representations attention mechanisms feature extraction multimodal segmentation embeddings interpretability

The Big Picture

Imagine a baby learning to make sense of the world. Long before anyone explains what a dog is, the infant has already made the connection: a furry creature appears, a bark rings out, and somewhere in that developing brain, a link forms. Later, when someone says “dog,” that same brain region lights up. The infant has learned, without instruction, to separate two very different kinds of knowledge: what something sounds like and what a word means.

This seemingly simple trick has eluded AI systems for years. Teaching an AI to link audio to video is one thing. Teaching it to understand that a bark is a physical event while the word “dog” is a symbolic reference, without ever being told the difference, is something else entirely.

Most modern audio-visual AI systems lump these two together, producing one-size-fits-all representations that can answer “did a dog appear in this video?” but can’t tell you where the dog is or which part of the audio referred to it.

Researchers from MIT, Oxford, Google, and Microsoft built DenseAV, a system that watches unlabeled videos and learns to locate sounds and match spoken words to objects in images. It spontaneously separates these two types of connection without ever being told to.

Key Insight: DenseAV discovers, entirely on its own, that spoken words and physical sounds represent different kinds of connections to the visual world, and it separates them into distinct processing channels using only the signal of paired video frames and audio.

How It Works

The architecture starts with two separate neural networks (one for audio, one for images) that process their respective signals into grids of features. Each feature is a compact numerical summary of what’s happening in a small patch of an image or a brief slice of audio. Rather than collapsing these grids into a single number, as most systems do, DenseAV keeps them dense, preserving fine-grained spatial and temporal detail.

Every image patch has its own feature summary. Every audio slice has its own. The system then computes a similarity volume: a structured table of match scores between every audio moment and every image patch.

Think of it as a correlation map. If the word “dog” is spoken at time t, and the dog occupies pixels in the upper-left corner, the match score between that audio moment and those image patches should be high. DenseAV is trained to make it so. No bounding boxes. No transcripts. Just the raw signal of co-occurrence.

Figure 1

The researchers then introduce a multi-head feature aggregation operator, a way of dividing processing into separate specialized channels, each independently learning what patterns to match:

  1. Dense feature maps are split into K separate “heads” (the paper explores K=1 and K=2).
  2. Each head independently computes its own AV similarity volume.
  3. Spatial dimensions are max-pooled, not average-pooled, forcing the system to identify the best-matching region rather than spreading credit across the whole image.
  4. Head dimensions are also max-pooled, allowing different heads to specialize.
  5. The audio time dimension is average-pooled to produce a final similarity score for training.

The max-pooling choice is subtle but it makes all the difference. Average pooling rewards a system for being vaguely right everywhere. Max pooling rewards it for being precisely right somewhere. That pressure is what drives DenseAV’s localization ability.

Figure 3

When trained on video datasets containing both narrated speech and ambient sound, something unexpected happened with the two-head model: one head spontaneously specialized in physical sounds (the bark of a dog), while the other specialized in spoken language (the word “dog”). No label ever told the system these were different. The distinction arose purely from what the training signal demanded; the only way to reduce errors was to treat these two kinds of audio differently.

Why It Matters

The gap DenseAV exposes in existing models is significant. ImageBind, Meta’s widely-used multimodal model, achieves impressive cross-modal retrieval scores (a measure of how well a system matches content across different media, like finding the right image for an audio clip). But when you visualize its local features, as shown in the paper’s qualitative comparisons, the alignment to image regions is diffuse at best.

Figure 2

DenseAV outperforms ImageBind on cross-modal retrieval using fewer than half the parameters. It also outperforms all prior work on new segmentation benchmarks the researchers contributed, tasks where you prompt the system with a spoken word or a sound and ask it to highlight the corresponding region of an image.

This matters beyond benchmarks. Applications from robotics to hearing aids to low-resource language documentation all require knowing not just that a connection exists between sound and image, but where and why.

That a machine learning system can discover, on its own, the distinction between “what something sounds like” and “what a word means” has deeper implications. It raises questions about what patterns in training data are doing the work, and whether the same pressures might explain how human infants develop the ability to connect their senses in the first place.

Bottom Line: DenseAV learns to localize spoken words and environmental sounds in images, and separates them into distinct representations, using nothing but unlabeled video. It outperforms systems with twice as many parameters and points toward a richer, grounded understanding of audio-visual AI.

IAIFI Research Highlights

Interdisciplinary Research Achievement
DenseAV directly mirrors a core puzzle in developmental neuroscience: how infants separate semantic language grounding from acoustic event localization. It provides a computational model that achieves the same feat from raw sensory data alone.
Impact on Artificial Intelligence
The multi-head dense similarity aggregation operator shows that contrastive objectives over *local* features, with max-pooling over spatial and head dimensions, unlock localization and disentanglement capabilities unavailable to global-representation architectures.
Impact on Fundamental Interactions
By learning how physical sound events differ structurally from symbolic language references, DenseAV probes the representational geometry underlying cross-modal perception, a question relevant to how information is encoded across interacting physical systems.
Outlook and References
Future directions include extending DenseAV to more heads for finer-grained audio category discovery and applying dense AV representations to low-resource language documentation; the work is available at https://aka.ms/denseav.

Original Paper Details

Title
Separating the "Chirp" from the "Chat": Self-supervised Visual Grounding of Sound and Language
arXiv ID
[arXiv:2406.05629](https://arxiv.org/abs/2406.05629)
Authors
["Mark Hamilton", "Andrew Zisserman", "John R. Hershey", "William T. Freeman"]
Abstract
We present DenseAV, a novel dual encoder grounding architecture that learns high-resolution, semantically meaningful, and audio-visually aligned features solely through watching videos. We show that DenseAV can discover the ``meaning'' of words and the ``location'' of sounds without explicit localization supervision. Furthermore, it automatically discovers and distinguishes between these two types of associations without supervision. We show that DenseAV's localization abilities arise from a new multi-head feature aggregation operator that directly compares dense image and audio representations for contrastive learning. In contrast, many other systems that learn ``global'' audio and video representations cannot localize words and sound. Finally, we contribute two new datasets to improve the evaluation of AV representations through speech and sound prompted semantic segmentation. On these and other datasets we show DenseAV dramatically outperforms the prior art on speech and sound prompted semantic segmentation. DenseAV outperforms the previous state-of-the-art, ImageBind, on cross-modal retrieval using fewer than half of the parameters. Project Page: \href{https://aka.ms/denseav}{https://aka.ms/denseav}