weipang142857's picture
Upload folder using huggingface_hub
f6cc031 verified
<Poster Width="1735" Height="1227">
<Panel left="-1" right="186" width="483" height="1035">
<Text>Motivation:</Text>
<Text>Robots face two challenges in natural environments:</Text>
<Text>● Underspecified goals: no human to specify exact</Text>
<Text>goal configuration</Text>
<Text>● Uncertain dynamics: Effects of robot's actions on</Text>
<Text>novel objects is uncertain</Text>
<Text>Approach:</Text>
<Text>● For underspecified goals:</Text>
<Text>• Pose task as a constrained optimization problem</Text>
<Text>over a set of reward or cost terms.</Text>
<Text>• Can be defined manually or modeled from human</Text>
<Text>● For uncertain dynamics:</Text>
<Text>• Quickly approximate dynamics for a set of actions</Text>
<Text>• Plan efficiently using sampling-based techniques</Text>
<Text>Our algorithm:</Text>
<Text>● Searches in object configuration space using</Text>
<Text>Rapidly-exploring Random Trees (RRT)</Text>
<Text>● Adds leaves to search tree by forward-simulating</Text>
<Text>the learned dynamics for each object-action pair</Text>
<Text>● Uses directGD heuristic to quickly search</Text>
<Text>optimization landscape</Text>
<Text>● Returns a plan from the starting state to the</Text>
<Text>most optimal reachable state, given cost function</Text>
<Figure left="40" right="895" width="403" height="316" no="1" OriWidth="0.400578" OriHeight="0.244861
" />
</Panel>
<Panel left="483" right="187" width="827" height="589">
<Text>Manipulation under uncertainty</Text>
<Text>Initial state</Text>
<Text>● Robot begins with a workspace containing </Text>
<Text>three unfamiliar objects</Text>
<Text>● Robot provided a cost function expressing </Text>
<Text>the following desiderata:</Text>
<Text>• Orthogonality</Text>
<Text>• Cicumscribed area</Text>
<Text>• Distance from edge of workspace</Text>
<Figure left="508" right="465" width="364" height="280" no="2" OriWidth="0.186127" OriHeight="0.181412
" />
<Text>Solution</Text>
<Text>● All objects pushed to orthonormal</Text>
<Text>orientations in the center of the workspace</Text>
<Text>● All paths free of collisions and redundant</Text>
<Text>actions</Text>
<Text>● Robot monitored error and replanned as</Text>
<Text>necessary</Text>
<Figure left="888" right="412" width="413" height="332" no="3" OriWidth="0.380347" OriHeight="0.183199
" />
</Panel>
<Panel left="486" right="778" width="828" height="441">
<Text>Generalization to other manipulation tasks</Text>
<Text>Appropriate for tasks naturally expressed as</Text>
<Text>optimization of a cost function:</Text>
<Text>● Arranging clutter on a surface</Text>
<Text>● Multiple object placement</Text>
<Text>● Table setting</Text>
<Figure left="520" right="995" width="328" height="216" no="4" OriWidth="0.176301" OriHeight="0.100536
" />
<Figure left="873" right="866" width="428" height="343" no="5" OriWidth="0.383237" OriHeight="0.208222
" />
</Panel>
<Panel left="1318" right="188" width="407" height="676">
<Text>Model Learning</Text>
<Text>Goal: discover the dynamics of each object</Text>
<Text>class over a set of action primitives</Text>
<Figure left="1342" right="296" width="366" height="191" no="6" OriWidth="0.30578" OriHeight="0.124218
" />
<Figure left="1335" right="500" width="380" height="195" no="7" OriWidth="0.744509" OriHeight="0.181859
" />
<Figure left="1366" right="709" width="315" height="143" no="8" OriWidth="0.406936" OriHeight="0.118409
" />
</Panel>
<Panel left="1318" right="873" width="406" height="346">
<Text>Advantages</Text>
<Text>Appropriate for tasks naturally expressed as</Text>
<Text>optimization of a cost function:</Text>
<Text>• Unlike conventional single-shot methods,</Text>
<Text>doesn't require user specified goals</Text>
<Text>• Always guaranteed to return reachable solution</Text>
<Text>• Favorable anytime characteristics</Text>
<Text>• Feasible for real-time planning in high DOF problems</Text>
<Text>Similar to Reinforcement Learning formalism, but</Text>
<Text>trades path optimality for realtime feasibility</Text>
<Text>• RL can require many full-passes through</Text>
<Text>configuration space to converge to optimal policy</Text>
<Text>• Handling continuous features requires discretization,</Text>
<Text>tiling, or other appoaches</Text>
<Text>• RL better suited for problems with sparse reward</Text>
<Text>landscape, but optimizations offer a gradient (like</Text>
<Text>shaping reward) which allows fast heuristic search</Text>
<Text>with RRT</Text>
</Panel>
</Poster>