{"title":"Web Log Analysis in Genealogy System","authors":"Xiaojian Liu, Yi Zhu, Shengwei Ji","doi":"10.1109/ICBK50248.2020.00081","DOIUrl":"https://doi.org/10.1109/ICBK50248.2020.00081","url":null,"abstract":"A large amount of log data is generated every moment in many real-world applications. System logs, mobile application logs, web logs etc., all of which contain rich information and can be analyzed for valuable information. At the same time, genealogy attracts more and more attention in recent decades, especially for Chinese. People utilize the genealogy system to seek roots, find source of family, build a family tree and so on. Due to the complexity of genealogy data and specific features of genealogy system, how to extract and analyze the valued genealogy information effectively is a huge challenge. Motivated by this, we propose a novel analysis method using web logs in genealogy system for drawing user profile and providing personalized recommendations. More specifically, firstly, all web logs are extracted from genealogy system based on the enhanced logs which record the behaviors and data of user. Secondly, the extracted logs are formatted for analysis, and we analyze the logs to extract various user-related attributes for drawing user profile. Finally, personalized recommendations like interested people and family tree are provided for each user in system. The high availability and effectiveness of our proposed method is verified on genealogy system-Huapu system.","PeriodicalId":432857,"journal":{"name":"2020 IEEE International Conference on Knowledge Graph (ICKG)","volume":"156 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125913873","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"EntityLDA: A Topic Model for Entity Retrieval on Knowledge Graph","authors":"Yu Hong, Suo Feng, Yanghua Xiao","doi":"10.1109/ICBK50248.2020.00062","DOIUrl":"https://doi.org/10.1109/ICBK50248.2020.00062","url":null,"abstract":"Encoding great aof information, knowledge graph (KG) has become a popular data source for information retrieval, especially the entity retrieval task. However, many online KGs include both structured triples and unstructured texts, which makes it difficult to represent entities in a unified form. Moreover, there is also a vocabulary gap between queries given by users and triples contained in KG. To solve these problems, we propose EntityLDA, a topic model which jointly models structured and unstructured parts of KG in order to get complete descriptions of entities. It also bridges the vocabulary gap between users and KG by connecting related words with shared topics. We further propose a retrieval solution based on EntityLDA to retrieve entities under different circumstances. Experimental results show that EntityLDA outperforms baselines in both quantity and quality.","PeriodicalId":432857,"journal":{"name":"2020 IEEE International Conference on Knowledge Graph (ICKG)","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127290302","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Integrating Bi-Dynamic Routing Capsule Network with Label-Constraint for Text classification","authors":"Xiang Guo, Youquan Wang, Kaiyuan Gao, Jie Cao, Haicheng Tao, Chaoyue Chen","doi":"10.1109/ICBK50248.2020.00011","DOIUrl":"https://doi.org/10.1109/ICBK50248.2020.00011","url":null,"abstract":"Neural-based text classification methods have attracted increasing attention in recent years. Unlike the standard text classification methods, neural-based text classification methods perform the representation operation and end-to-end learning on the text data. Many useful insights can be derived from neural based text classifiers as demonstrated by an ever-growing body of work focused on text mining. However, in the real-world, text can be both complex and noisy which can pose a problem for effective text classification. An effective way to deal with this issue is to incorporate self-attention and capsule networks into text mining solutions. In this paper, we propose a Bi-dynamic routing Capsule Network with Label-constraint (BCNL) model for text classification, which moves beyond the limitations of previous methods by automatically learning the task-relevant and label-relevant words of text. Specifically, we use a Bi-LSTM and self-attention with position encoder network to learn text embeddings. Meanwhile, we propose a bi-dynamic routing capsule network with label-constraint to adjust the category distribute of text capsules. Through extensive experiments on four datasets, we observe that our method outperforms state-of-the-art baseline methods.","PeriodicalId":432857,"journal":{"name":"2020 IEEE International Conference on Knowledge Graph (ICKG)","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126563255","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Approximate Single-Peakedness in Large Elections","authors":"Zhihuai Chen, Q. Li, Xiaoming Sun, Lirong Xia, Jialin Zhang","doi":"10.1109/ICBK50248.2020.00068","DOIUrl":"https://doi.org/10.1109/ICBK50248.2020.00068","url":null,"abstract":"Single-peaked preferences are a natural way to avoid paradoxes and impossibility theorems in social choice and have recently been involved in the study of various computational aspects of social choice. Since strict single-peakedness is hard to achieve in practice, approximate single-peakedness appears more appropriate and is gaining popularity. In this paper, we study approximate single-peakedness of large, randomly-generated profiles. We focused on characterizing the asymptotically optimal social axis, which is asymptotically consistent with most agents’ preferences generated from a statistical model. We characterize all asymptotically optimal social axes under the Mallows model for two case: the case where the dispersion parameter $varphi$ is close to 0, and the case where $varphi$ is close to 1. We also design an algorithm to help characterize all asymptotically optimal social axes for all $varphi$ when the number of alternative is no more than 10. These results help us understand the structure of approximate single-peakedness in large elections.","PeriodicalId":432857,"journal":{"name":"2020 IEEE International Conference on Knowledge Graph (ICKG)","volume":"47 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127644266","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Hybrid Framework of Convolution and Recurrent Neural Networks for Text Classification","authors":"Shengfei Lyu, Jiaqi Liu","doi":"10.1109/ICBK50248.2020.00052","DOIUrl":"https://doi.org/10.1109/ICBK50248.2020.00052","url":null,"abstract":"Convolutional neural network (CNN) and recurrent neural network (RNN) are two popular architectures used in text classification. Traditional methods to combine the strengths of the two networks rely on streamlining them or concatenating features extracted from them. In this paper, we propose a novel method to keep the strengths of the two networks to a great extent. In the proposed model, a convolutional neural network is applied to learn a 2D weight matrix where each row reflects the importance of each word from different aspects. Meanwhile, we use a bidirectional RNN to process each word and employ a neural tensor layer that fuses forward and backward hidden states to get word representations. In the end, the weight matrix and word representations are combined to obtain the representation in a 2D matrix form for the text. We carry out experiments on a number of datasets for text classification. The experimental results confirm the effectiveness of the proposed method.","PeriodicalId":432857,"journal":{"name":"2020 IEEE International Conference on Knowledge Graph (ICKG)","volume":"89 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123250880","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Improving Document-level Relation Extraction via Contextualizing Mention Representations andWeighting Mention Pairs","authors":"Ping Jiang, Xian-Ling Mao, Bin-Bin Bian, Heyan Huang","doi":"10.1109/ICBK50248.2020.00051","DOIUrl":"https://doi.org/10.1109/ICBK50248.2020.00051","url":null,"abstract":"Document-level relation extraction (RE) has attracted considerable attention, because a large number of relational facts are expressed in multiple sentences. Recently, encoder-aggregator based models have become promising for document-level RE. However, these models have two shortcomings: (i) they cannot obtain contextualized representations of a mention by low computational cost, when the mention is involved in different entity pairs; (ii) they ignore the different weights for the mention pairs of a target entity pair. To tackle the above two problems, in this paper, we propose a novel encoder-attender-aggregator model, which introduces two attenders between the encoder and aggregator. Specifically, a mutual attender is first employed on the selected head and tail mentions to efficiently produce contextualized mention representations. Then, an integration attender is utilized to weight the mention pairs of a target entity pair. Extensive experiments on two document-level RE datasets show that the proposed model performs better than the state-of-the-art baselines. Our codes are publicly available at “https://github.com/nefujiangping/EncAttAgg”.","PeriodicalId":432857,"journal":{"name":"2020 IEEE International Conference on Knowledge Graph (ICKG)","volume":"263 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131414033","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Prediction of Financial Big Data Stock Trends Based on Attention Mechanism","authors":"Jiannan Chen, Junping Du, Zhe Xue, Feifei Kou","doi":"10.1109/ICBK50248.2020.00031","DOIUrl":"https://doi.org/10.1109/ICBK50248.2020.00031","url":null,"abstract":"Stock trend prediction has always been the focus of research in the field of financial big data. Stock data is complex nonlinear data, while stock price is changing over time. Based on the characteristics of stock data, this paper proposes a financial big data stock trend prediction algorithm based on attention mechanism (STPA). We adopt Bidirectional Gated Recurrent Unit (BGRU) and attention mechanism to capture the long-term dependence of data on time series. The attention mechanism is used to analyze the weight of the impact of data from different time periods on the trend prediction results, thereby reducing the error of stock data change trend prediction and improving the accuracy of trend prediction. We select the daily closing price data of 10 stocks for model training and performance evaluation. Experimental results demonstrate that the proposed method STPA achieves higher precision, recall rate and F1-Score in predicting stock change trends than the other methods. Compared with mainstream methods, STPA improves the precision by 4%, improves recall by 2.5%, and improves F1-Score by 3.2%.","PeriodicalId":432857,"journal":{"name":"2020 IEEE International Conference on Knowledge Graph (ICKG)","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114699256","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Relation Extraction with Proactive Domain Adaptation Strategy","authors":"Lingfeng Zhong, Yi Zhu","doi":"10.1109/ICBK50248.2020.00069","DOIUrl":"https://doi.org/10.1109/ICBK50248.2020.00069","url":null,"abstract":"Relation extraction is an important information extraction task in many Natural Language Processing (NLP) applications, such as automatic knowledge graph construction, question answering, sentiment analysis, etc. However, relation extraction suffers from inappropriate associations between entities when the background knowledge of corpus is insufficiency. Despite the preprocessed external word vector bases can ease this problem, how to find a single word vector base as domain knowledge that contains all the required knowledge features is a huge challenge, and relation extraction with background knowledge is still open to further optimization. To address this problem, in this paper, we propose Relation Extraction method with Proactive Domain Adaptation Strategy (REPDAS for short) to introduce more knowledge features from different knowledge bases. More specifically, firstly, a convolutional network with a parameter-sharing layer is introduced for relation extraction, and word seeds that are important to relational feature exploitation are proactively picked by an attention mechanism during training. Secondly, the proactively-chosen word seeds and the previous parameter-sharing layer are utilized to establish a map between different domains. Our proposed method selectively avails both background knowledge and contextual features for relation extraction by incorporating the convolutional neural network with the proactively domain adaptation strategy. Experiments show that our method effectively enhances the performance of relation extraction compared with other baselines.","PeriodicalId":432857,"journal":{"name":"2020 IEEE International Conference on Knowledge Graph (ICKG)","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114989078","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Online Markov Blanket Discovery With Streaming Features","authors":"Dianlong You, Ruiqi Li, Miaomiao Sun, Xinju Ou, Shunpan Liang, Fuyong Yuan","doi":"10.1109/ICBK50248.2020.00023","DOIUrl":"https://doi.org/10.1109/ICBK50248.2020.00023","url":null,"abstract":"The Markov blanket (MB) in Bayesian networks has attracted much attention since the MB of a target attribute (T) is the minimal feature subset with maximum prediction ability for classification. Nevertheless, traditional MB discovery methods such as IAMB, HITON-MB, and MMMB are not suitable for streaming features. Meanwhile, online feature selection with streaming features (OSFSF) methods such as Alpha-investing and SAOLA, focus only on the relevance and ignores causality of the features, they cannot mine the MB, or only find the parents and children (PC) such as OSFS. Therefore, these methods have weaker interpretability and do not have sufficient prediction accuracy. We propose a novel algorithm for online markov blanket discovery with streaming features to tackle abovementioned issues, named OMBSF. OMBSF finds the MB of T containing parents, children, and spouses, discards false positives from the PC and spouse sets online, distinguishes between the PC and spouse in real-time. An empirical study demonstrates that OMBSF finds a more accurate MB when the volume of features is high and it significantly improves the prediction accuracy than other algorithms. Moreover, OMBSF obtains a bigger feature subset than that obtained by OSFS, demonstrating that OMBSF can identify the spouse in the MB that are not identified using OSFS.","PeriodicalId":432857,"journal":{"name":"2020 IEEE International Conference on Knowledge Graph (ICKG)","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123354580","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Multi-task Classification Model Based On Multi-modal Glioma Data","authors":"Jialun Li, Yuanyuan Jin, Hao Yu, Xiaoling Wang, Qiyuan Zhuang, Liang Chen","doi":"10.1109/ICBK50248.2020.00033","DOIUrl":"https://doi.org/10.1109/ICBK50248.2020.00033","url":null,"abstract":"Glioma is a common disease. According to relevant medical research, there is a specific relationship between the appearance of glioma and the genotype of isocitrate dehydrogenase-1 (IDHI). It is also affected by 1p/19q chromosome deletion status. This study uses deep learning techniques to explore the relationship among glioma morphology, IDH1 genotypes and 1p/19q chromosomes based on multi-modal glioma data. We train CNN to obtain the intensity, location and shape of glioma according to MRI images. Taking the features of glioma as input, we use XGBoost to classify the IDH1 genotype and and SVM to classify 1p/19q chromosome status. We find that processing the brain MRI images through CNN can accurately obtain some medical feature information of the glioma, and the accuracy rate of the model is above 0.8. When classifying IDH1 genotype and 1p/19q chromosome status based on these features, we find that the image features of gliomas are more closely related to the IDH1 genotype than to the 1p/19q chromosome status.","PeriodicalId":432857,"journal":{"name":"2020 IEEE International Conference on Knowledge Graph (ICKG)","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129603984","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}