1. 📘 Topic and Domain: The paper focuses on real-time Vision-Language-Action (VLA) models for robotic manipulation, specifically addressing reaction latency in flow-based action chunking policies.
2. 💡 Previous Research and New Ideas: The paper builds on existing flow-based VLAs (π0.5, X-VLA) and asynchronous inference methods, proposing a novel Horizon-Aware Schedule that prioritizes immediate actions during flow sampling to achieve 10× faster reaction times without architectural changes.
3. ❓ Problem: The paper solves the reaction latency bottleneck in action chunking VLA policies, where constant timestep schedules force completion of all sampling steps before any movement can start, limiting real-time responsiveness in dynamic tasks.
4. 🛠️ Methods: The paper introduces FASTER (Fast Action Sampling for ImmediaTE Reaction) using a Horizon-Aware Schedule that adaptively allocates sampling steps across the action chunk, enabling single-step generation of immediate actions while maintaining long-horizon trajectory quality, coupled with a streaming client-server interface.
5. 📊 Results and Evaluation: FASTER achieves 10× acceleration in Time to First Action (TTFA) compared to baselines, demonstrates superior performance in real-world tasks including table tennis (0.80 vs 0.20 score on RTX 4090), and maintains competitive performance on simulation benchmarks (96.5% on LIBERO, 4.292 on CALVIN).