<span id="fpn9h"><noframes id="fpn9h"><span id="fpn9h"></span>
<span id="fpn9h"><noframes id="fpn9h">
<th id="fpn9h"></th>
<strike id="fpn9h"><noframes id="fpn9h"><strike id="fpn9h"></strike>
<th id="fpn9h"><noframes id="fpn9h">
<span id="fpn9h"><video id="fpn9h"></video></span>
<ruby id="fpn9h"></ruby>
<strike id="fpn9h"><noframes id="fpn9h"><span id="fpn9h"></span>

異構三機器人協同搬運的高柔順性研究

Research on High Flexibility of Heterogeneous three-robot Collaborative Handling

  • 摘要: 針對異構三機器人系統的協同搬運柔順性問題,提出基于近端策略優化(Proximal Policy Optimization)的強化學習控制方法。在CoppeliaSim機器人仿真器中建立了異構三機器人協同搬運的仿真環境,分別開展了力控制與強化學習控制的對比仿真。仿真結果表明:強化學習控制下,物體質心的軌跡誤差在Z方向上最優,僅為力控制的4.7%,機器人2的末端速度變化和其典型關節的角速度變化更為平滑。采用sim2real的方法,將兩種控制方法部署到三機器人協同搬運實驗中。實驗結果表明:強化學習控制下,Z方向的物體軌跡跟蹤誤差同樣最優,僅為力控制的5.4%。機器人2在X方向上的速度變化僅為力控制的20.7%,其典型關節展現出更好的柔順性,角速度變化僅為力控制下的35.2%。仿真與實驗結果表明:強化學習的控制效果更優,也具備從仿真到現實遷移的可行性。

     

    Abstract: Aiming at the flexibility issue in the collaborative handling of the heterogeneous three-robot system, a reinforcement learning control strategy based on Proximal Policy Optimization (PPO) is proposed. A simulation environment for the collaborative handling of the heterogeneous three robots is established in the CoppeliaSim robot simulator, and the simulations of force control and reinforcement learning control are performed respectively. The comparison of simulation results show that under the reinforcement learning control, the trajectory error of the mass center of the handled object is the smallest in the Z direction, which is only 4.7% of that under the force control. The end effector velocity change of robot 2 and the angular velocity change of its typical joint are much smoother. Through the sim2real solution, both control methods are deployed in the three-robot collaborative handling experiment. The experimental results indicate that under the reinforcement learning control, the object trajectory tracking error in the Z direction is also the smallest, only 5.4% of that under the force control. The velocity change of robot in the X direction is only 20.7% of that under the force control, and its typical joint 2 reveals better flexibility, with the angular velocity change being only 35.2% of that under the force control. The simulation and experimental results demonstrate that the reinforcement learning method is provided with better performance, and it also has the feasibility of being transferred from simulation to reality.

     

/

返回文章
返回
<span id="fpn9h"><noframes id="fpn9h"><span id="fpn9h"></span>
<span id="fpn9h"><noframes id="fpn9h">
<th id="fpn9h"></th>
<strike id="fpn9h"><noframes id="fpn9h"><strike id="fpn9h"></strike>
<th id="fpn9h"><noframes id="fpn9h">
<span id="fpn9h"><video id="fpn9h"></video></span>
<ruby id="fpn9h"></ruby>
<strike id="fpn9h"><noframes id="fpn9h"><span id="fpn9h"></span>
www.77susu.com