mateoguaman/vamos_10pct_gpt5_mini_cocoqa_localized_narratives Viewer • Updated Sep 14, 2025 • 1.29M
VAMOS: A Hierarchical Vision-Language-Action Model for Capab Collection This collection contains VLM planner checkpoints, affordance module checkpoints for Spot and HOUND, training datasets, and a demo • 7 items • Updated Oct 27, 2025 • 2
VAMOS: A Hierarchical Vision-Language-Action Model for Capab Collection This collection contains VLM planner checkpoints, affordance module checkpoints for Spot and HOUND, training datasets, and a demo • 7 items • Updated Oct 27, 2025 • 2