节点文献

基于知识表达的迁移学习研究

Transfer Learning Research Based on Knowledge Representation

【作者】 张倩

【导师】 李明;

【作者基本信息】 中国矿业大学 , 控制理论与控制工程, 2013, 博士

【摘要】 随着计算机信息技术的飞速发展,从海量数据中挖掘有用的信息并加以利用已经成为当前的研究热点。在数据挖掘中,传统的机器学习假设训练数据与测试数据具有相同的分布,而实际应用中往往并不能满足这一假设,即使勉强满足也大大降低了数据分析的效率。迁移学习方法能够有效地在相似的领域或任务之间进行信息的共享和迁移,使传统的从零开始的学习变成可积累的学习,具有成本低,效率高等优点,近几年成为数据挖掘和机器学习的关注焦点。迁移学习最大的特点就是利用相关领域的知识来帮助完成目标领域中的学习任务。但是,不同的知识表达方式直接影响迁移学习的效果。本文从知识的表达方式入手,对知识表达与迁移方法的结合方式展开研究,主要工作如下:首先,针对源领域数据和目标领域数据分布类似的情况,提出一种基于多源动态TrAdaBoost的实例迁移学习方法。该方法考虑多个源领域知识,使得目标任务的学习可以充分利用所有源领域信息;每次训练候选分类器时,所有源领域样本都参与学习,可以获得有利于目标任务学习的有用信息,从而避免负迁移的产生。理论分析验证了所提算法较单源迁移的优势,以及加入动态因子改善了源权重收敛导致的权重熵由源样本转移到目标样本的问题。实验结果表明,此算法在提高识别率和分类精度等方面的优势。其次,针对源领域数据和目标领域数据分布差别较大的情况,提出一种基于协方差矩阵的特征迁移学习方法。该方法采用相似学习估计个体参数的协方差对,从数据间特征联系出发,通过构造半正定规划来组合这些估计值并学习当前任务的先验值,构建多变量高斯先验的协方差矩阵,进而对未标记的数据进行有效预测。实验结果表明,所提方法利用少量的源任务即可得到较好的迁移学习结果。再次,针对共享知识是一些基函数的参数模型或先验分布的情况,提出一种基于层次贝叶斯的参数迁移学习方法。该方法定义任务相关性的层次概念,对标准Dirichlet过程模型进行扩展,建立多任务环境下的贝叶斯强化学习模型;有效学习类别数量、类型的推断过程以及用于探索新任务时的类别结构,从而更快地学习更多的知识。实验结果表明,提出的方法可以加快新领域中最优策略的收敛速度。最后,针对共享知识是规则、结构和逻辑等关联规则的情况,提出一种基于马尔可夫逻辑网的关联规则迁移学习方法。该方法首先利用伪对数似然函数,将源领域中马尔可夫逻辑网表示的知识迁移到目标领域中,建立两个领域之间的关联;通过对源领域进行自诊断、结构更新和目标领域搜索新子句,来优化映射得到的结构,进而适应目标领域的学习。实验结果表明,算法成功地映射了迁移知识,提高了学习模型的精确度。全文针对目前迁移学习研究中存在的部分问题,以如何利用源领域和目标领域间共享的不同知识表达方式为出发点,以提高目标领域的学习效率为目的,以不同的迁移方法为基本手段,对如何构造迁移环境和如何建立迁移环境下的有效迁移模型进行深入研究,提出了解决办法。实验结果验证了上述方法的可行性和有效性。论文研究成果丰富了迁移学习理论,为其在数据挖掘中的应用提供了有益的指导。

【Abstract】 Along with the rapid development of computer information technology, to mineuseful information from mass data and put them into use have become a currentresearch hot spot. In data mining, the assumption for the traditional machine learningis that training data and test data have the same distribution. While in the practicalapplication, this assumption cannot be often met. To meet this assumption barely, theefficiency of data analysis has to be reduced. With the use of different fields ofknowledge for target task learning, transfer learning can transfer and share theinformation between similar domains or tasks, making the traditional learning fromscratch an addable one, which implies that the learning efficiency is higher and thecost is lower, so that it has become the focus in the field of data mining and machinelearning in recent years.The most remarkable characteristic of transfer learning is that it can employ theknowledge in relative domains to help perform the learning tasks in the domain of thetarget. However, different ways of knowledge expression directly affect the effect oftransfer learning. Beginning with the ways of knowledge expression, aiming at thecombination modes of knowledge expression and transfer methods, this paper unfoldsits study mainly as follows:Firstly, specific to the situation that the shared knowledge in the domains of thesource and the target are sample data with similar distribution, an instance transferlearning method based on multi-sources dynamic TrAdaBoost is put forward.Integrated with the knowledge in multiple source domains, this method makes thetarget task learning the one that is able to make good use of the information of allsource domains. Whenever candidate classifiers are trained, all the samples in allsource domains are involved in learning, and the information conducive to target tasklearning can be obtained, so that negative transfer can be avoided. The theoreticalanalysis suggests that the given algorithm better than single source transfer, by meansof adding the dynamic factor, this algorithm improves the defect that weight entropyto drift from source to target instances. The experimental results support that the givenalgorithm has the advantage of improving the recognition rate and classificationaccuracy.Secondly, specific to the situation that shared knowledge in the domains of thesource and the target are sample data with different distribution, a feature transfer learning based on covariance matrix is proposed. This method employs similarlearning to estimate the covariance pairs of individual parameters. Starting from thecharacteristic relation between data, by means of constructing semidefinite program,the estimated values are combined, the priori values of the current tasks are emulated,and the priori multivariate gaussian covariance matrix is automatically built in order toeffectively predict the unlabelled data. The experimental results show that, with theuse of a small amount of the source tasks, good transfer learning results can beachieved by the given method.Thirdly, specific to the situation that shared knowledge are the parametric modelsor the priori distributions of some base functions, a parameter transfer learningmethod based on hierarchical bayes is brought forward. With the use of this method,the hierarchy concept of task dependency is defined, the standard Dirichlet processmodel is developed, and then the bayes learning model is built under multi-taskenvironment. The number of categories, the inference process of types and thecategory structure used for exploring a new task are automatically studied, thus moreknowledge are learned quickly. The experimental results show that the proposedmethod can speed up the rate of convergence of the optimal strategy in the newdomain.Finally, specific to the situation that shared knowledge are associated knowledgecontaining rules, structure and logic, an rule-based transfer learning methodassociated with Markov logic network is presented. Having applied this method, bymeans of pseudo log-likelihood function, the knowledge in the source domainexpressed in Markov logic network are transferred into the target domain while thelink between the two domains is established. By means of the self diagnosis andstructure update in the source domain and the new clause surf in the target domain,the mapped structure is optimized so that it can be adapt to the learning in the targetdomain. The experimental results show that the given algorithm successfully maps thetransferred knowledge, and improves the precision of the learning model.Aiming at some existing problems in the current study on transfer learning, withhow to utilize the different knowledge expressions shared in both source and targetdomains as the starting point, with improving the learning efficiency in the targetdomain as the purpose, with different transfer methods as the basic means, with anin-depth study on how to construct the transfer environment and how to build theeffective transfer model under the transfer environment, some solutions are given in the full text. The related simulation results have verified the feasibility and availabilityof the given solutions. The research results in this paper have enriched transferlearning theories, which provide a helpful guidance for complex data mining.

节点文献中: 

本文链接的文献网络图示:

本文的引文网络