\n
\n\n \n \n \n \n \n Bayesian Optimization for Whole-Body Control of High-Degree-of-Freedom Robots Through Reduction of Dimensionality.\n \n \n \n\n\n \n Yuan, K.; Chatzinikolaidis, I.; and Li, Z.\n\n\n \n\n\n\n
IEEE Robotics and Automation Letters, 4(3). 2019.\n
\n\n
\n\n
\n\n
\n\n \n\n \n \n doi\n \n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n \n \n 1 download\n \n \n\n \n \n \n \n \n \n \n\n \n \n \n \n \n \n \n \n \n \n \n\n\n\n
\n
@article{\n title = {Bayesian Optimization for Whole-Body Control of High-Degree-of-Freedom Robots Through Reduction of Dimensionality},\n type = {article},\n year = {2019},\n keywords = {Optimization and optimal control,humanoid and bipedal locomotion,humanoid robots,legged robots},\n volume = {4},\n id = {5df89826-3412-3eac-953e-6315a585a899},\n created = {2021-04-21T21:00:15.548Z},\n file_attached = {false},\n profile_id = {fcec045e-0e8d-3ec2-90c3-ece3d1622c91},\n last_modified = {2021-04-21T21:00:15.548Z},\n read = {false},\n starred = {false},\n authored = {true},\n confirmed = {false},\n hidden = {false},\n private_publication = {true},\n abstract = {This letter aims to achieve automatic tuning of optimal parameters for whole-body control algorithms to achieve the best performance of high-DoF robots. Typically, the control parameters at a scale up to hundreds are often hand-tuned yielding sub-optimal performance. Bayesian optimization (BO) can be an option to automatically find optimal parameters. However, for high-dimensional problems, BO is often infeasible in realistic settings as we studied in this letter. Moreover, the data is too little to perform dimensionality reduction techniques, such as principal component analysis or partial least square. We hereby propose an alternating BO algorithm that iteratively learns the parameters of sub-spaces from the whole high-dimensional parametric space through interactive trials, resulting in sample efficiency and fast convergence. Furthermore, for the balancing and locomotion control of humanoids, we developed techniques of dimensionality reduction combined with the proposed ABO approach that demonstrated optimal parameters for robust whole-body control.},\n bibtype = {article},\n author = {Yuan, K. and Chatzinikolaidis, I. and Li, Z.},\n doi = {10.1109/LRA.2019.2901308},\n journal = {IEEE Robotics and Automation Letters},\n number = {3}\n}
\n
\n\n\n
\n This letter aims to achieve automatic tuning of optimal parameters for whole-body control algorithms to achieve the best performance of high-DoF robots. Typically, the control parameters at a scale up to hundreds are often hand-tuned yielding sub-optimal performance. Bayesian optimization (BO) can be an option to automatically find optimal parameters. However, for high-dimensional problems, BO is often infeasible in realistic settings as we studied in this letter. Moreover, the data is too little to perform dimensionality reduction techniques, such as principal component analysis or partial least square. We hereby propose an alternating BO algorithm that iteratively learns the parameters of sub-spaces from the whole high-dimensional parametric space through interactive trials, resulting in sample efficiency and fast convergence. Furthermore, for the balancing and locomotion control of humanoids, we developed techniques of dimensionality reduction combined with the proposed ABO approach that demonstrated optimal parameters for robust whole-body control.\n
\n\n\n
\n
\n\n \n \n \n \n \n Learning Whole-Body Motor Skills for Humanoids.\n \n \n \n\n\n \n Yang, C.; Yuan, K.; Merkt, W.; Komura, T.; Vijayakumar, S.; and Li, Z.\n\n\n \n\n\n\n In
IEEE-RAS International Conference on Humanoid Robots, volume 2018-Novem, 2019. \n
\n\n
\n\n
\n\n
\n\n \n\n \n \n doi\n \n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@inproceedings{\n title = {Learning Whole-Body Motor Skills for Humanoids},\n type = {inproceedings},\n year = {2019},\n volume = {2018-Novem},\n id = {3631b402-eeff-31d7-88c4-75e92b58fd95},\n created = {2021-04-21T21:00:16.118Z},\n file_attached = {false},\n profile_id = {fcec045e-0e8d-3ec2-90c3-ece3d1622c91},\n last_modified = {2021-04-21T21:00:16.118Z},\n read = {false},\n starred = {false},\n authored = {true},\n confirmed = {false},\n hidden = {false},\n private_publication = {true},\n abstract = {This paper presents a hierarchical framework for Deep Reinforcement Learning that acquires motor skills for a variety of push recovery and balancing behaviors, i.e., ankle, hip, foot tilting, and stepping strategies. The policy is trained in a physics simulator with realistic setting of robot model and low-level impedance control that are easy to transfer the learned skills to real robots. The advantage over traditional methods is the integration of high-level planner and feedback control all in one single coherent policy network, which is generic for learning versatile balancing and recovery motions against unknown perturbations at arbitrary locations (e.g., legs, torso). Furthermore, the proposed framework allows the policy to be learned quickly by many state-of-the-art learning algorithms. By comparing our learned results to studies of preprogrammed, special-purpose controllers in the literature, self-learned skills are comparable in terms of disturbance rejection but with additional advantages of producing a wide range of adaptive, versatile and robust behaviors.},\n bibtype = {inproceedings},\n author = {Yang, C. and Yuan, K. and Merkt, W. and Komura, T. and Vijayakumar, S. and Li, Z.},\n doi = {10.1109/HUMANOIDS.2018.8625045},\n booktitle = {IEEE-RAS International Conference on Humanoid Robots}\n}
\n
\n\n\n
\n This paper presents a hierarchical framework for Deep Reinforcement Learning that acquires motor skills for a variety of push recovery and balancing behaviors, i.e., ankle, hip, foot tilting, and stepping strategies. The policy is trained in a physics simulator with realistic setting of robot model and low-level impedance control that are easy to transfer the learned skills to real robots. The advantage over traditional methods is the integration of high-level planner and feedback control all in one single coherent policy network, which is generic for learning versatile balancing and recovery motions against unknown perturbations at arbitrary locations (e.g., legs, torso). Furthermore, the proposed framework allows the policy to be learned quickly by many state-of-the-art learning algorithms. By comparing our learned results to studies of preprogrammed, special-purpose controllers in the literature, self-learned skills are comparable in terms of disturbance rejection but with additional advantages of producing a wide range of adaptive, versatile and robust behaviors.\n
\n\n\n