AlanaVLM: A Multimodal Embodied AI Foundation Model for Egocentric Video UnderstandingJan 1, 2024ยทAlessandro Suglia,Claudio Greco,Katie Baker,Jose L. Part,Ioannis Papaioannou,Arash Eshghi,Ioannis Konstas,Oliver Lemonยท 0 min read Cite DOI URLTypeJournal articlePublicationCoRRLast updated on Jan 1, 2024 ← AlanaVLM: A Multimodal Embodied AI Foundation Model for Egocentric Video Understanding Jan 1, 2024CROPE: Evaluating In-Context Adaptation of Vision and Language Models to Culture-Specific Concepts Jan 1, 2024 →