Multimodal Deep Learning: Not A Review (Keep Updating)
The competition in the field of computer vision is so fierce that I have to extend my silly research into the area of multimodality.
Just kidding! Actually, I do believe that multimodal research is important. On the one hand, perception is inherently multimodal and leveraging multimodal information may aid in making more accurate decisions. For instance, humans perceive the world through eyes, ears, noses, etc. These sense organs as a whole play a key role in making us one of the most successful species. On the other hand, modern models, particularly large language models (LLMs), offer a unified interface for different modalities, enabling the development of a truly multimodal system.
I’ve put together a brief review here, based on an online lecture I watched before. But I have to say that I couldn’t find the link at present.
Note: This brief review focuses on task performance and training efficiency, rather than inference efficiency, and does not consider deployment.
Introduction
What is multimodal
Merriam-webster’s Dictionary of College English is unclear:
multimodal: having or involving several modes, modalities, or maxima
A previous review [Liang et al., 2022] gives another definition:
A modality refers to a way in which a natural phenomenon is perceived or expressed. Multimodal refers to situations where multiple modalities are involved.
Why multimodal
- Real: Human perception of the external world is multimodal.
- Practical: The internet and various applications are primarily multimodal.
- Effective: Multimodal data contains richer information and is often easier to obtain (available textual information is nearly exhausted).
Multimodal research challenges
I recommend everyone to read this review [Liang et al., 2022], which goes beyond specific techniques to model and understand multimodal problems at a higher level.
Example Applications
Take the applications between text and image as an example.
- Retrieval (Image<>Text)
- Caption (Image->Text)
- Generation (Text->Image)
- Visual Question Answering (Image + Text -> Text)
- Multimodal Classification (Image + Text -> Label)
- Better Understanding & Generation (Image + Text -> Label/Text)
Model Evolution
Multimodal learning is not confined to specific tasks. Actually, it spans multiple technological eras and exhibits many common characteristics across these eras. Therefore, this section will primarily follow the timeline of technological eras to trace the evolution of multimodal model structures, thereby revealing the development trajectory of multimodal learning. In particular, vision and language are the most extensively studied tasks, and this section will focus on detailed discussions through multimodal tasks between images and text.
Early Models
Since the mid-20th century, multimodal learning has remained a major research area in artificial intelligence. Given that deep learning has become the prevailing method for realizing artificial intelligence today, this section begins with an introduction to research related to neural networks.
Typical Early Studies
In the early 2000s, fundamental modern deep learning components, such as MLP and word embedding, were already in use. By the middle of the decade, deep learning models (e.g., CNN, RNN, GNN) had already achieved state-of-the-art performance in various tasks. Additionally, generative models like GANs also appeared. At the end of the decode, the widespread application of Transformer provided a similar paradigm for different tasks, driving the overall development of the field of artificial intelligence. Until today, deep learning technology is still on the rise.
References
[Liang et al., 2022] Liang, Paul Pu, Amir Zadeh, and Louis-Philippe Morency. “Foundations and recent trends in multimodal machine learning: Principles, challenges, and open questions." arXiv preprint arXiv:2209.03430 (2022).
[Weston et al., 2011] Weston, Jason, Samy Bengio, and Nicolas Usunier. “Wsabie: Scaling up to large vocabulary image annotation." Twenty-Second International Joint Conference on Artificial Intelligence. 2011.