Project: Embodied AI for Multimodal Semantic Understanding
Project, UTS AI Institute, Sydney, Australia
This project integrates multiple modalities for comprehensive AI perception and interaction in embodied agents. By combining vision, language, and physical interaction capabilities, we aim to create AI systems that can understand and navigate the physical world with human-like comprehension.