05.12.2012 Views

russell l. tedrake - MIT Computer Science and Artificial Intelligence ...

russell l. tedrake - MIT Computer Science and Artificial Intelligence ...

russell l. tedrake - MIT Computer Science and Artificial Intelligence ...

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

[15] Ian R. Manchester, Mark M. Tobenkin, Michael Levashov, <strong>and</strong> Russ Tedrake. Regions of attraction for<br />

hybrid limit cycles of walking robots. Proceedings of the 18th IFAC World Congress, extended version<br />

available online: arXiv:1010.2247 [math.OC], 2011.<br />

[16] John Roberts, Ian Manchester, <strong>and</strong> Russ Tedrake. Feedback controller parameterizations for reinforcement<br />

learning. In Proceedings of the 2011 IEEE Symposium on Adaptive Dynamic Programming <strong>and</strong><br />

Reinforcement Learning (ADPRL), 2011.<br />

[17] Ian R. Manchester, Uwe Mettin, Fumiya Iida, <strong>and</strong> Russ Tedrake. Stable dynamic walking over uneven<br />

terrain. The International Journal of Robotics Research (IJRR), 30(3):265–279, January 24 2011.<br />

[18] Alex<strong>and</strong>er Shkolnik, Michael Levashov, Ian R. Manchester, <strong>and</strong> Russ Tedrake. Bounding on rough<br />

terrain with the littledog robot. The International Journal of Robotics Research (IJRR), 30(2):192–<br />

215, Feb 2011.<br />

[19] Jan Peters, Russ Tedrake, Nicholas Roy, <strong>and</strong> Jun Morimoto. Robot learning. In Claude Sammut <strong>and</strong><br />

Geoffrey I. Webb, editors, Encyclopedia of Machine Learning, pages 865–9. Springer, 2011.<br />

[20] Mark Tobenkin, Ian R. Manchester, Jennifer Wang, Alex Megretski, <strong>and</strong> Russ Tedrake. Convex optimization<br />

in identification of stable non-linear state space models. In Proceedings of the 49th IEEE<br />

Conference on Decision <strong>and</strong> Control (CDC 2010), extended version available online: arXiv:1009.1670<br />

[math.OC], Dec 2010.<br />

[21] R. Platt, R. Tedrake, L.P. Kaelbling, <strong>and</strong> T. Lozano-Perez. Belief space planning assuming maximum<br />

likelihood observations. In Proceedings of Robotics: <strong>Science</strong> <strong>and</strong> Systems, 2010.<br />

[22] Russ Tedrake, Ian R. Manchester, Mark M. Tobenkin, <strong>and</strong> John W. Roberts. LQR-Trees: Feedback motion<br />

planning via sums of squares verification. International Journal of Robotics Research, 29:1038–<br />

1052, July 2010.<br />

[23] Philipp Reist <strong>and</strong> Russ Tedrake. Simulation-based LQR-trees with input <strong>and</strong> state constraints. In<br />

Proceedings of the International Conference on Robotics <strong>and</strong> Automation (ICRA), 2010.<br />

[24] Elena Glassman <strong>and</strong> Russ Tedrake. A quadratic regulator-based heuristic for rapidly exploring state<br />

space. In Proceedings of the International Conference on Robotics <strong>and</strong> Automation (ICRA), 2010.<br />

[25] Fumiya Iida <strong>and</strong> Russ Tedrake. Minimalistic control of biped walking in rough terrain. Autonomous<br />

Robots, OnlineFirst:355–368, January 7 2010.<br />

[26] Jan Peters, Jun Morimoto, Russ Tedrake, <strong>and</strong> Nicholas Roy. Robot learning [TC Spotlight]. Robotics<br />

<strong>and</strong> Automation Magazine, IEEE, 16(3):19–20, September 2009.<br />

[27] John W. Roberts, Jun Zhang, <strong>and</strong> Russ Tedrake. Motor learning at intermediate reynolds number:<br />

Experiments with policy gradient on the flapping flight of a rigid wing. In From Motor to Interaction<br />

Learning in Robots. Springer, 2009.<br />

[28] Russ Tedrake. LQR-Trees: Feedback motion planning on sparse r<strong>and</strong>omized trees. In Proceedings of<br />

Robotics: <strong>Science</strong> <strong>and</strong> Systems (RSS), page 8, 2009.<br />

Russell Tedrake 3 of 14

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!