节点文献
机器人鲁棒学习控制
Robust Learning Control of Robots
【作者】 顾武军;
【导师】 王聪;
【作者基本信息】 华南理工大学 , 控制理论与控制工程, 2010, 硕士
【摘要】 近二十多年来,以神经网络、模糊逻辑和进化算法为代表的人工智能理论与方法开始被运用到机器人系统的控制当中。神经网络具有通用的逼近能力、分布式的结构,控制过程中可以不需要系统的数学模型知识等优点,并且具有与其它控制理论如自适应控制,变结构控制等很强的融合能力,因此神经网络被广泛应用于含有未知动力学的机器人系统的控制。稳定自适应控制是近年来基于神经网络的非线性系统自适应控制方法的研究主流,在保证稳定性方面有了越来越严格的理论证明。但是自适应神经网络控制中的一个不足或局限就是网络需要反复训练,不像人脑一样具有一次学习的能力,而且不能将学习到的知识重新利用。同样基于自适应神经网络的机器人控制虽然取得了很大的成绩,但是对于系统的未知动力学在控制任务完成后仍然是未知的。研究者对于神经网络有没有真正学习到系统的未知动力学未进行探索,每次即使重复相同的控制任务都要进行网络的重新训练,而这一训练过程是耗时耗能的,所以很有必要进一步探讨神经网络在稳定的自适应闭环控制过程中学习未知动力学知识的能力,并且期望能将学习到的动力学知识运用到相同或者相似的控制任务中,避免网络的重复训练,同时使得自适应神经网络控制器能根据经验累积知识,改善控制性能。根据确定学习原理,对于一个合理设计的自适应神经网络控制器,在跟踪周期或者类周期的参考轨迹时,其径向基函数网络的回归向量满足部分持续激励条件,进一步分析可得到网络估计权值的部分收敛,根据径向基函数网络的局部特性,最终系统的未知非线性能够由一个常值神经网络逼近。本文利用确定学习的最新成果,提出了含有未知动力学和干扰的机器人鲁棒学习控制策略。在稳定的闭环控制过程中,所设计的鲁棒自适应神经网络控制器在跟踪周期或者类周期的参考轨迹时,回归向量满足部分持续激励条件,通过分析由闭环系统推导而来的线性时变系统的指数稳定性,可得到网络估计权值的部分收敛,从而神经网络可以学习到机器人闭环系统未知的动力学,并将学到的动力学作为经验知识以常值网络权值的形式表示。通过学习,使我们了解系统未知动力学模型特性成为可能,并且在下次重复相同或者相似的控制任务时,控制器可以调用以往所学动力学知识用于控制并获得更好的控制性能。该策略避免了耗时耗能的神经网络重新训练过程,使得机器人具有真正意义上的从经历中获取知识,表达知识,并将学到的知识再利用的能力。
【Abstract】 In the past decades, artificial intelligence theory and methods such as neural networks, fuzzy logics and evolution algorithms have been used to control robot system. Neural networks have lots of advantages including universal approximation abilities, distributed architecture, accurate system mathematic model is not needed in control process and powerful ability of working with other control theory such as adaptive control, variable structure control, so neural networks have been widely applied to the control of robot system with unknown dynamics.In recent years, stable adaptive control has been the main trend of neural network based adaptive control of nonlinear system, and there is stricter theoretical proof on system stability. But one of the shortness of adaptive neural networks is that the network need to be trained repeatedly, do not have the ability of learning as human brains and reusing the learned knowledge. Great achievements have been made toward adaptive neural networks based robot control, however, the system unknown dynamics are still unknown when the control task is finished. The researchers have not explored whether the neural network have learned the unknown dynamics, even for the same control task the neural networks have to be trained repeatedly, which is time and energy wasted. So it is meaningful to further investigate the neural networks’ability of learning the unknown dynamics in the stable adaptive closed loop control process. The learned knowledge is expected to be reused in the same or similar control tasks so that the repeated training phase can be avoided. Moreover, the adaptive neural network controller is expected to accumulate knowledge by experiences so that better control performance can be achieved.According to deterministic learning theory, for an appropriately designed adaptive neural network controller, the sub-vector of the radial basis function networks’regression vector satisfy the persistent exciting condition when tracking period or period-like desired trajectory, the estimate errors of neural network weights are proved to be convergent to a small neighborhood of zero, and ultimately the unknown the nonlinearity could be learned by a constant neural network due to the local property of radial basis function networks.In this thesis, by using the achievements of deterministic learning theory, robust learning control scheme is presented for robots with unknown dynamics and disturbances. In the stable closed loop control process, the regression vector of the designed robust adaptive neural network controller satisfies partial persistent exciting condition. By analyzing the linear time varying system obtained from the closed loop system, partial neural network weights convergence can be achieved. The unknown dynamics of closed loop robot system can be learned by neural network, and saved as experience knowledge in the form of constant neural weights. The learning enables us to understand the underlying characteristic of the unknown system dynamics. When repeat the same or similar control task, the controller can also reuse the learned dynamic knowledge and better control performance can be achieved with little efforts. Time and energy wasting repeated training phase can be avoided through the proposed scheme, both of the true learning and the reusing of learned knowledge are realized.
【Key words】 adaptive neural network; persistent exciting; deterministic learning; robust adaptive control; robust learning control;