{"title":"用有限数量的标记训练样本学习马尔可夫逻辑网络","authors":"Tak-Lam Wong","doi":"10.3233/KES-140289","DOIUrl":null,"url":null,"abstract":"Markov Logic Networks (MLN) is a unified framework integrating first-order logic and probabilistic inference. Most existing methods of MLN learning are supervised approaches requiring a large amount of training examples, leading to a substantial amount of human effort for preparing these training examples. To reduce such human effort, we have developed a semi-supervised framework for learning an MLN, in particular structure learning of MLN, from a set of unlabeled data and a limited number of labeled training examples. To achieve this, we aim at maximizing the expected pseudo-log-likelihood function of the observation from the set of unlabeled data, instead of maximizing the pseudo-log-likelihood function of the labeled training examples, which is commonly used in supervised learning of MLN. To evaluate our proposed method, we have conducted experiments on two different datasets and the empirical results demonstrate that our framework is effective, outperforming existing approach which considers labeled training examples alone.","PeriodicalId":210048,"journal":{"name":"Int. J. Knowl. Based Intell. Eng. Syst.","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2014-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":"{\"title\":\"Learning Markov logic networks with limited number of labeled training examples\",\"authors\":\"Tak-Lam Wong\",\"doi\":\"10.3233/KES-140289\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Markov Logic Networks (MLN) is a unified framework integrating first-order logic and probabilistic inference. Most existing methods of MLN learning are supervised approaches requiring a large amount of training examples, leading to a substantial amount of human effort for preparing these training examples. To reduce such human effort, we have developed a semi-supervised framework for learning an MLN, in particular structure learning of MLN, from a set of unlabeled data and a limited number of labeled training examples. To achieve this, we aim at maximizing the expected pseudo-log-likelihood function of the observation from the set of unlabeled data, instead of maximizing the pseudo-log-likelihood function of the labeled training examples, which is commonly used in supervised learning of MLN. To evaluate our proposed method, we have conducted experiments on two different datasets and the empirical results demonstrate that our framework is effective, outperforming existing approach which considers labeled training examples alone.\",\"PeriodicalId\":210048,\"journal\":{\"name\":\"Int. J. Knowl. Based Intell. Eng. Syst.\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2014-04-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"3\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Int. J. Knowl. Based Intell. Eng. Syst.\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.3233/KES-140289\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Int. J. Knowl. Based Intell. Eng. Syst.","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.3233/KES-140289","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Learning Markov logic networks with limited number of labeled training examples
Markov Logic Networks (MLN) is a unified framework integrating first-order logic and probabilistic inference. Most existing methods of MLN learning are supervised approaches requiring a large amount of training examples, leading to a substantial amount of human effort for preparing these training examples. To reduce such human effort, we have developed a semi-supervised framework for learning an MLN, in particular structure learning of MLN, from a set of unlabeled data and a limited number of labeled training examples. To achieve this, we aim at maximizing the expected pseudo-log-likelihood function of the observation from the set of unlabeled data, instead of maximizing the pseudo-log-likelihood function of the labeled training examples, which is commonly used in supervised learning of MLN. To evaluate our proposed method, we have conducted experiments on two different datasets and the empirical results demonstrate that our framework is effective, outperforming existing approach which considers labeled training examples alone.