Skip to content

Commit

Permalink
Grammar and minor fixes
Browse files Browse the repository at this point in the history
  • Loading branch information
abhaybd committed Apr 23, 2024
1 parent 7f66e1c commit 586b603
Showing 1 changed file with 14 additions and 7 deletions.
21 changes: 14 additions & 7 deletions index.html
Original file line number Diff line number Diff line change
Expand Up @@ -214,7 +214,15 @@ <h3 class="title publication-title">Data Efficient Behavior Cloning for Fine Man
<h2 class="title is-2">Abstract</h2>
<div class="content has-text-justified">
<p>
We stuy imitation learning to teach robots from expert demonstrations. During robots' execution, compounding errors from hardware noise and external disturbances, coupled with incomplete data coverage, can drive the agent into unfamiliar states and cause unpredictable behaviors. To address this challenge, we propose a framework, CCIL: Continuity-based data augmentation for Corrective Imitation Learning. It leverages the local continuity inherent in dynamic systems to synthesize corrective labels. CCIL learns a dynamics model from the expert data and uses it to generate labels guiding the agent back to expert states. Our approach makes minimal assumptions, requiring neither expert re-labeling nor ground truth dynamics models. By exploiting local continuity, we derive provable bounds on the errors of the synthesized labels. Through evaluations across diverse robotic domains in simulation and the real world, we demonstrate CCIL's effectiveness in improving imitation learning performance.
We study imitation learning to teach robots from expert demonstrations. During robots' execution, compounding errors
from hardware noise and external disturbances, coupled with incomplete data coverage, can drive the agent into
unfamiliar states and cause unpredictable behavior. To address this challenge, we propose a framework, CCIL:
Continuity-based data augmentation for Corrective Imitation Learning. It leverages the local continuity inherent in
dynamic systems to synthesize corrective labels. CCIL learns a dynamics model from the expert data and uses it to
generate labels guiding the agent back to expert states. Our approach makes minimal assumptions, requiring neither
expert re-labeling nor ground truth dynamics models. By exploiting local continuity, we derive provable bounds on the
errors of the synthesized labels. Through evaluations across diverse robotic domains in simulation and the real world,
we demonstrate CCIL's effectiveness in improving imitation learning performance.
</p>
</div>
</div>
Expand Down Expand Up @@ -462,7 +470,7 @@ <h4 class="title is-4">CCIL improves imitation learning, especially in low-data
<div class="columns is-centered">
<div class="column is-9 has-text-centered insight">
<p>
CCIL can bring in prominent performance boost in low-data regimes compared to using standard behavior cloning, showcasing
CCIL can yield a prominent performance boost in low-data regimes compared to using standard behavior cloning, showcasing
its data efficiency and robustness.
</p>
</div>
Expand All @@ -479,12 +487,11 @@ <h4 class="title is-4">CCIL's Robustness to Lipschitz Constraint</h4>
<div class="columns is-centered">
<div class="column is-9 has-text-centered insight">
<p>
CCIL makes a critical assumption that the system dynamics contain local continuity. In practice, however, its application is relatively insensitive to the hyper-parameter choice of Lipschitz constraint in learning the dynamics model. As long as we filter generated labels using appropriate label error threshold, CCIL could yield a significant performance boost.
CCIL makes a critical assumption that the system dynamics contain local continuity. In practice, however, its
application is relatively insensitive to the hyper-parameter choice of Lipschitz constraint in learning the dynamics
model. As long as we filter generated labels using appropriate label error threshold, CCIL could yield a significant
performance boost.
</p>
<!--<p>
This is corroborated by the observation that different Lipschitz constraints yield similar distributions of local Lipschitz constants,
indicating that the model is able to capture the environment's inherent continuity without strict regularization.
</p>-->
</div>
</div>
</div>
Expand Down

0 comments on commit 586b603

Please sign in to comment.