Action Chunking for Imitation Learning

Enhancing VQ-BET performance via action chunking, hyperparameter sweeps, and temporal moving averages at NYU GRAIL Lab.

Action Chunking for Imitation Learning

Overview

Enhancing VQ-BET performance via action chunking, hyperparameter sweeps, and temporal moving averages at NYU GRAIL Lab.

Project Overview

Conducted at the GRAIL Lab at NYU Courant under Professor Lerrel Pinto, this research focused on enhancing the performance of a vanilla Vector Quantized Behavioral Transformer (VQ-BET) model. I implemented action chunking—a technique where the model predicts a sequence of future actions rather than a single step—to significantly improve inference smoothness and mitigate compounding execution errors.

Technical Details

The experimentation phase involved extensive hyperparameter sweeps across blend and buffer decay factors, temporal moving averages, and chunk lengths. By testing models with various time skips (e.g., short, zero, and 7-step long skips), the project identified the optimal configurations for stable, long-horizon imitation learning tasks.

Robotics Imitation Learning Machine Learning