Multimodal AI
Learning across modalities — vision, language, audio, time-series, and beyond. Open exploration of how AI can see, hear, read, and reason across any domain simultaneously.
Landscape
The central question: how do we build models that perceive, represent, and reason across multiple modalities simultaneously — the way humans do, effortlessly? Vision, language, audio, time-series, 3D geometry, sensor data, tabular signals — these are not separate problems. They are facets of the same underlying challenge: grounding intelligence in the full richness of the world.
This area is intentionally kept open. It's an exploration zone — not constrained to a single application domain or modality pair. The framing: multimodal AI is the frontier where representation learning, generative modeling, and foundation models converge.
Sub-areas
- Multimodal representation learning — joint embedding spaces where different modalities align: contrastive (CLIP, ALIGN), masked reconstruction (BEiT-3, data2vec), and fusion-based approaches
- Multimodal generative learning — generating across modalities: text-to-image (Stable Diffusion, DALL-E, Imagen), text-to-audio, text-to-video, and any-to-any generation (UnifiedIO, Unified-IO 2)
- Vision-language models (VLMs) — grounding language in visual perception: LLaVA, InstructBLIP, GPT-4V, Gemini; reasoning about images with language
- Audio and speech — speech representation (wav2vec 2.0, HuBERT), audio-language models, music understanding
- Multimodal for science and industry — applying multimodal AI to non-standard domains: medical imaging + clinical notes, remote sensing + text, time-series + language, financial signals + news
- Embodied multimodal AI — agents that perceive and act in physical environments: RT-2, PaLM-E, SayCan; the intersection with robotics and world models
- Cross-modal transfer and zero-shot generalization — how does knowledge learned in one modality transfer to another? When does it generalize and when does it fail?
Landmark papers
- Learning Transferable Visual Models From Natural Language Supervision (CLIP) — Radford et al., 2021. The pivot point for vision-language alignment at scale.
- Flamingo: a Visual Language Model for Few-Shot Learning — Alayrac et al., 2022. Cross-modal few-shot learning with frozen LLMs.
- ImageBind: One Embedding Space To Bind Them All — Girdhar et al., 2023. Binding 6 modalities in a single embedding space using image as a hub.
- data2vec: A General Framework for Self-Supervised Learning in Speech, Vision and Language — Baevski et al., 2022. A unified self-supervised objective across modalities.
- Scaling Autoregressive Multi-Modal Models: Pretraining and Instruction Tuning (Gemini, GPT-4V era). The convergence of language models with vision.
- LLaVA: Large Language and Vision Assistant — Liu et al., 2023. Simple visual instruction tuning; strong baseline for VLMs.
- Unified-IO: A Unified Model for Vision, Language, and Structured Data — Lu et al., 2022. The "any-to-any" framing.
Key figures
Alec Radford (CLIP, GPT), Alexei Efros (vision foundations), Douwe Kiela (multimodal research), Christoph Feichtenhofer (video), Dhruv Batra (embodied AI + VQA), Serge Belongie (vision), Jitendra Malik (vision + embodied). The generative side: Robin Rombach (Stable Diffusion), Srinivas Narayanan.
Open Problems
-
Modality gap and alignment. CLIP-style models learn aligned representations, but the embeddings of different modalities often occupy different regions of the shared space — the "modality gap." True alignment (not just correlation) is still unsolved.
-
Compositional multimodal reasoning. Models can describe images or answer visual questions, but compositional reasoning — "if the red object were blue, would the scene still be symmetric?" — remains brittle. The failure modes are revealing about what's actually learned.
-
Efficient cross-modal transfer. Full pretraining on all modalities is expensive. How do you take a strong unimodal model (a language model, a vision encoder) and extend it to new modalities efficiently? Adapter-based approaches (LLaVA) are a step but not a complete answer.
-
Multimodal for non-visual domains. Most multimodal research is vision-language. The space of other useful modality combinations — audio-vision, time-series-language, table-language, sensor-language — is far less explored despite high practical value.
-
Evaluation that measures understanding. Existing benchmarks (VQA, COCO captions) are noisy proxies. A model can pass many benchmarks via superficial patterns without genuine cross-modal understanding. What would a rigorous evaluation look like?
-
Generation fidelity vs. grounding. Generative models produce impressive outputs but hallucinate. The tension between fluency (generation quality) and faithfulness (factual grounding to input) is fundamental and unresolved.
-
Any-to-any generation at scale. Can one model, trained jointly, generate any output modality from any input modality with no degradation relative to specialized models? The gap between specialized and universal models is still large.
Questions & Ideas
- Does CLIP-style contrastive learning actually produce grounded representations, or just correlated ones? How would you test the difference?
- Is there a natural "multimodal prior" for pretraining that's more data-efficient than paired data at scale?
- How much of multimodal generalization is inherited from the language model backbone vs. learned from vision-language pairs?
- What would a truly modality-agnostic tokenizer look like? Can you represent an image, a spectrogram, a time-series, and a sentence in a unified token vocabulary without domain-specific encoding?
- Are there modality combinations that are particularly "easy" for representation alignment and why? (vision-language seems easier than vision-audio — what makes some pairs more alignable?)
- What's the right architecture for real-time multimodal perception in embodied agents — where latency matters and you can't run a 70B model?
- For tabular + language: can a vision-language-style contrastive objective work for table-description pairs? What would "visual" mean for tabular data?
- Hypothesis: the best multimodal representations emerge not from direct cross-modal contrastive learning but from a world model that must predict one modality from another autoregressively.
My Take
This section evolves as thinking develops.
The most interesting frontier in multimodal AI is not the mainstream vision-language benchmark race. It's the question of what multimodal learning tells us about intelligence itself. A system that truly understands across modalities isn't just doing multimodal lookup — it's building a coherent model of the world where different sensory streams are integrated into a unified representation of reality.
My interest is in two directions:
Multimodal representation learning for non-standard domains. Most of the field's energy goes into vision-language. But the same principles — contrastive alignment, masked reconstruction, self-supervised grounding — apply to other modality pairs that matter more in science and industry: sensor data + language, time-series + reports, molecular structure + text. The methodology exists; the application is open territory.
The connection to world models. I think multimodal AI and world models are converging. A good world model needs to integrate visual, proprioceptive, linguistic, and temporal signals to predict how the environment will evolve. The "multimodal" framing in computer vision and the "world model" framing in RL are studying the same underlying problem from different angles.
Journal
2026-02-28 — Adding this as a sixth research area. The motivation is open exploration — I don't want to constrain multimodal work to one application domain. The area serves as a capture zone for anything multimodal: VLM experiments, audio-vision work, cross-modal representation learning, or applying AI to unusual domain combinations.
The near-term plan: work through CLIP and ImageBind carefully (understand the geometry of the shared embedding space), then run a small experiment on cross-modal transfer to a non-vision domain. The tabular + language direction from the causality area is the natural bridge.