ACoT-VLA | arxiv 2026.01.16 | Paper Reading
ACoT-VLA: Action Chain-of-Thought for Vision-Language-Action Models
ACoT-VLA | arxiv 2026.01.16 | Paper Reading
ACoT-VLA: Action Chain-of-Thought for Vision-Language-Action Models
InternVLA-A1 | arxiv 2026.01.05 | Paper Reading
InternVLA-A1: Unifying Understanding, Generation and Action for Robotic Manipulation
PointWorld | arxiv 2026.01.07 | Paper Reading
PointWorld: Scaling 3D World Models for In-The-Wild Robotic Manipulation
Learning to Remember: Exploring Multimodal Memory Mechanisms in Long Video Understanding
keywords: Memory, Long Video Understanding, VLA
MemoryVLA | ICLR 2026 | Paper Reading
MEMORYVLA: PERCEPTUAL-COGNITIVE MEMORY IN VISION-LANGUAGE-ACTION MODELS FOR ROBOTIC MANIPULATION
$\pi_{0.5}$ | arxiv 2025.04.22 | Paper Reading
InternVLA M1 | arxiv 2025.10.15 | Paper Reading
InternVLA-M1: A Spatially Guided Vision-Language-Action Framework for Generalist Robot Policy
SP-VLA | arxiv 2025.10.03 | Paper Reading
SP-VLA: A JOINT MODEL SCHEDULING AND TOKEN PRUNING APPROACH FOR VLA MODEL ACCELERATION
OpenVLA-OFT | RSS 2025 | Paper Reading
Fine-Tuning Vision-Language-Action Models: Optimizing Speed and Success
OpenVLA | CoRL 2024 | Paper Reading
OpenVLA: An Open-Source Vision-Language-Action Model