Blog link: https://www.pi.website/research/real_time_chunking

Problem with VLAs are that they’re slow to inference, and this could mean that robot takes non-optimal actions and mess up a real world task (eg. spill coffee in your lap)

For VLAs you need a system that lets the model think about it’s future actions while executing a previous one. Action Chunking (inferencing 50 actions for every 1 inference call) is a good starting point but when you generate new chunks old chunks might not “agree” with the new ones