COMP6411C: Advanced Topics in Multimodal Machine Learning
Spring 2024, Computer Science and Engineering, Hong Kong University of Science and Technology
Instructor: Long Chen
Class Time: Monday 13:30pm - 14:50pm, Friday 9:00am - 10:20am (Room 2504)
Email: longchen@ust.hk
(For course-related queries, please use the subject starting from [COMP6411C]
)
Office Hour: Friday 13:00 pm - 14:00 pm (Room 3508, ZOOM: 986 8247 0232 (Passcode: 6411C))
Teaching Assistant: Mr. Chaoyang Zhu (cy.zhu@connect.ust.hk
)
TA Office Hour: Monday 14:00 pm - 15:00 pm (Room 4204)
For those who have enrolled the COMP6411C course, if you want to get the recorded videos for absent classes, you can direct sent emails to the TA.
Student Presentation Schedule: Google Docs
Course Description: This course provides a comprehensive introduction to recent advances in multimodal machine learning, with a focus on vision-language research. Major topics include multimodal translation, multimodal reasoning, multimodal alignment, multimodal information extraction, and recent deep learning techniques in multimodal research (such as graph convolution network, Transformer architecture, deep reinforcement learning, and causal inference). The course structure will primarily consist of instructor presentation, student presentation, in-class discussion, and a course final project.
Course Objectives: After completion of this course, students will understand mainstream multimodal topics and tasks, and develop their critical thinking and problem solving, such as identifying and explaining the state-of-the-art approaches for multimodal applications.
Pre-requisite: Basic understanding of probability and linear algebra is required. Familiarity or experience with machine learning (especially deep learning) and computer vision basics are preferred.
Grading scheme:
- Class attendance and in-class discussion: 20%
- Project presentation: 30%
- Final project report: 50%
Reference books/materials:
- Conferences: Proceedings of CVPR/ICCV/ECCV, ICLR/ICML/NeurIPS, ACL/EMNLP/ACM Multimedia
- Book: Goodfellow, Ian, Yoshua Bengio, and Aaron Courville. Deep learning. MIT press, 2016.
Student Presentation Slides
-
1) Connect, Collapse, Corrupt: Learning Cross-Modal Tasks with Uni-Modal Data. In ICLR, 2024
-
5) 3D-LLM: Injecting the 3D World into Large Language Models. In NeurIPS, 2023.
-
7) Link-Context Learning for Multimodal LLMs. In CVPR, 2024.
-
8) Towards 3D Molecule-Text Interpretation in Language Models. In ICLR, 2024.
-
9) Balanced Multimodal Learning via On-the-fly Gradient Modulation. In CVPR, 2022.
-
11) ShapeGPT: 3D Shape Generation with A Unified Multi-modal Language Model. In arXiv, 2023.
-
12) A Symbolic Character-Aware Model for Solving Geometry Problems. In ACM MM, 2023.
-
14) PromptMRG: Diagnosis-Driven Prompts for Medical Report Generation. In arXiv, 2023.
-
16) MoE-LLaVA: Mixture of Experts for Large Vision-Language Models. In arXiv, 2024.
-
21) Making Large Multimodal Models Understand Arbitrary Visual Prompts. In CVPR, 2024.
-
23) Multi-granularity Correspondence Learning from Long-tern Noisy Videos. In ICLR, 2024.
-
25) Training Language Models to Follow Instructions with Human Feedback. In NeurIPS, 2022.
-
27) V-IRL: Grounding Virtual Intelligence in Real Life. In arXiv, 2024.
-
28) PMR: Prototypical Modal Rebalance for Multimodal Learning. In CVPR, 2023.
-
34) EgoVLPv2: Egocentric Video-Language Pre-training with Fusion in the Backbone, In ICCV, 2023.
-
37) Multimodal prompting with missing modalities for visual recognition. In CVPR, 2023.
-
38) Multimodal Pathway: Improve Transformers with Irrelevant Data from Other Modalities. In CVPR, 2024.
-
39) GraphTranslator: Aligning Graph Model to Large Language Model for Open-Ended Tasks. In WWW, 2024.
-
40) Improving multimodal datasets with image captioning. In NeurIPS, 2023.
-
42) HyperDiffusion: Generating Implicit Neural Fields with Weight-Space Diffusion. In arXiv, 2023.
Syllabus / Schedule
Course overview
Key research problems in multimodal learning
Multimodal learning tasks
Image representation
Convolutional neural networks
Image-based scene understanding
Video representation
Word representation
Recurrent neural networks
Revisit long video encoding
Encoder-decoder framework
Attention mechanism
Image captioning
New directions on image captioning
Diffusion mdels
Controllable text-to-image generation
Image editing
Concept customization
VQA tasks
VQA methods
Multi-step VQA
Robust VQA
Video retrieval
Temporal grounding
Weakly-supervised temporal grounding
Referring expression comprehension and phrase grounding
Referring expression segmentation
Weakly-supervised visual grounding
Scene graphs
Event Extraction
Event-event Relation
Graph Basics
Graph Convolution Network
General Perspective on GNN
Self-Attention and Transformer
Vision Transformer (ViT)
Policy gradient
Q-Learning
Backdoor adjustment
Frontdoor adjustment
Conterfactual thinking
BERT
Image-Text Pretreaining
Unified Pretraining
Video-Text Pretraining
Multi-channel Videos
Adapter, LoRA
Prompt tuning
Multimodal LLMs
LLM-enhanced multimodal learning
Acknowledgements
This course was inspired by and/or uses reserouces from the following courses:
MultiModal Machine Learning by Louis-Philippe Morency, Carnegie Mellon University, Fall 2023.
Advanced Topics in MultiModal Machine Learning by Louis-Philippe Morency, Carnegie Mellon University, Spring 2023.
Advances in Computer Vision by Bill Freeman, MIT, Spring, 2023.
Deep Learning for Computer Vision by Fei-Fei Li, Stanford University, Spring 2023.
Natural Language Processing with Deep Learning by Christopher Manning, Stanford University, Winter 2023.
Deep Learning for Computer Visionby Justin Johnson, University of Michigan, Winter 2022.