<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>SO-101 | Home - Jyothi Swaroop</title><link>https://kjyothiswaroop.github.io/tag/so-101/</link><atom:link href="https://kjyothiswaroop.github.io/tag/so-101/index.xml" rel="self" type="application/rss+xml"/><description>SO-101</description><generator>Hugo Blox Builder (https://hugoblox.com)</generator><language>en-us</language><lastBuildDate>Sun, 01 Mar 2026 00:00:00 +0000</lastBuildDate><image><url>https://kjyothiswaroop.github.io/media/icon_hue06d895dbab0d0c9b72b1a0534685c49_26160_512x512_fill_lanczos_center_3.png</url><title>SO-101</title><link>https://kjyothiswaroop.github.io/tag/so-101/</link></image><item><title>Deformable Object Manipulation</title><link>https://kjyothiswaroop.github.io/project/lehome-challenge/</link><pubDate>Sun, 01 Mar 2026 00:00:00 +0000</pubDate><guid>https://kjyothiswaroop.github.io/project/lehome-challenge/</guid><description>&lt;hr>
&lt;h2 id="overview">Overview&lt;/h2>
&lt;p>The &lt;strong>leHome Challenge&lt;/strong> (ICRA 2026) is a &lt;strong>Deformable BiManipulation Challenge&lt;/strong> focused on folding laundry using a bimanual &lt;strong>SO-101&lt;/strong> robot. Teams must train policies capable of manipulating deformable objects — one of the hardest open problems in robot learning due to the near-infinite configuration space of cloth.&lt;/p>
&lt;p>We finished &lt;strong>54th out of 230+ teams&lt;/strong>.&lt;/p>
&lt;hr>
&lt;h2 id="data-collection">Data Collection&lt;/h2>
&lt;p>Automatic data collection was attempted by extracting &lt;strong>privileged information&lt;/strong> about the cloth mesh directly from the &lt;strong>Isaac Sim&lt;/strong> scene and passing goal poses to &lt;strong>cuRobo&lt;/strong> for bimanual IK solving and motion planning. In practice this pipeline was too brittle — cuRobo IK failures and sim–real cloth geometry mismatches meant trajectories rarely transferred to usable demonstrations.&lt;/p>
&lt;p>We fell back to &lt;strong>manual teleoperation&lt;/strong> on the physical SO-101 arms, then significantly expanded the dataset through &lt;strong>data augmentation&lt;/strong>: applying colour-jitter, brightness, contrast, and blur filters to camera observations to improve policy generalisation across lighting conditions.&lt;/p>
&lt;hr>
&lt;h2 id="policy">Policy&lt;/h2>
&lt;p>Five policy architectures were trained and evaluated:&lt;/p>
&lt;table>
&lt;thead>
&lt;tr>
&lt;th>Policy&lt;/th>
&lt;th>Notes&lt;/th>
&lt;/tr>
&lt;/thead>
&lt;tbody>
&lt;tr>
&lt;td>&lt;strong>ACT&lt;/strong>&lt;/td>
&lt;td>Action Chunking with Transformers&lt;/td>
&lt;/tr>
&lt;tr>
&lt;td>&lt;strong>Diffusion Policy&lt;/strong>&lt;/td>
&lt;td>Denoising diffusion over action sequences&lt;/td>
&lt;/tr>
&lt;tr>
&lt;td>&lt;strong>SmolVLA&lt;/strong>&lt;/td>
&lt;td>Vision-Language-Action model — &lt;strong>best performance&lt;/strong>&lt;/td>
&lt;/tr>
&lt;tr>
&lt;td>&lt;strong>LingBotVLA&lt;/strong>&lt;/td>
&lt;td>Language-conditioned VLA&lt;/td>
&lt;/tr>
&lt;/tbody>
&lt;/table>
&lt;p>&lt;strong>SmolVLA&lt;/strong> achieved the highest task success rate across our evaluations. All policies were trained using the &lt;strong>LeRobot&lt;/strong> framework with &lt;strong>PyTorch&lt;/strong>.&lt;/p>
&lt;hr>
&lt;h2 id="results">Results&lt;/h2>
&lt;p>Ranked &lt;strong>54th / 230+ teams&lt;/strong> at the leHome Challenge (ICRA 2026).&lt;/p></description></item></channel></rss>