Multimodal - 2024-02

Publish Date Title Authors PDF Translate Read Code
2024-02-29 Panda-70M: Captioning 70M Videos with Multiple Cross-Modality Teachers Tsai-Shien Chen et.al. 2402.19479 translate read null
2024-02-29 FATE in MMLA: A Student-Centred Exploration of Fairness, Accountability, Transparency, and Ethics in Multimodal Learning Analytics Yueqiao Jin et.al. 2402.19071 translate read null
2024-02-28 Grounding Language Models for Visual Entity Recognition Zilin Xiao et.al. 2402.18695 translate read link
2024-02-28 Multimodal Learning To Improve Cardiac Late Mechanical Activation Detection From Cine MR Images Jiarui Xing et.al. 2402.18507 translate read null
2024-02-28 DecisionNCE: Embodied Multimodal Representations via Implicit Preference Learning Jianxiong Li et.al. 2402.18137 translate read null
2024-02-27 Multimodal Learned Sparse Retrieval with Probabilistic Expansion Control Thong Nguyen et.al. 2402.17535 translate read link
2024-02-27 Curriculum Learning Meets Directed Acyclic Graph for Multimodal Emotion Recognition Cam-Van Thi Nguyen et.al. 2402.17269 translate read null
2024-02-26 GROUNDHOG: Grounding Large Language Models to Holistic Segmentation Yichi Zhang et.al. 2402.16846 translate read null
2024-02-26 Gradient-Guided Modality Decoupling for Missing-Modality Robustness Hao Wang et.al. 2402.16318 translate read null
2024-02-24 FedMM: Federated Multi-Modal Learning with Modality Heterogeneity in Computational Pathology Yuanzhe Peng et.al. 2402.15858 translate read null
2024-02-20 GRAFFORD: A Benchmark Dataset for Testing the Knowledge of Object Affordances of Language and Vision Models Sayantan Adak et.al. 2402.12881 translate read link
2024-02-19 Multimodal Emotion Recognition from Raw Audio with Sinc-convolution Xiaohui Zhang et.al. 2402.11954 translate read null
2024-02-18 Efficient Multimodal Learning from Data-centric Perspective Muyang He et.al. 2402.11530 translate read link

(<a href=../Multimodal.md>back to Multimodal</a>)