Proceedings of the 6th International Conference on Natural Language Processing and Knowledge Engineering(NLPKE-2010)最新文献

筛选
英文 中文
Are we waves or are we particles? A new insight into deep semantics in natural language processing 我们是波还是粒子?自然语言处理中深层语义的新见解
Svetlana Machova, J. Klecková
{"title":"Are we waves or are we particles? A new insight into deep semantics in natural language processing","authors":"Svetlana Machova, J. Klecková","doi":"10.1109/NLPKE.2010.5587805","DOIUrl":"https://doi.org/10.1109/NLPKE.2010.5587805","url":null,"abstract":"This paper brings conceptually new, empirically based scientific approach to a deeper understanding of human mind cognition, language acquisition, modularity of language and language origin itself. The research presented provides an interactive multilingual associative experiment as an attempt to map the Cognitive Semantic Space: (CSSES) and its basic frames of the Essential Self in the Czech language, collects and compares it to the CSSES of conceptual language view in Czech, Russian, English and potentially in other languages. We attempt to merge cognitive metaphor theory with psycholinguistics and psychoanalysis applying associative experiment methodology on the Essential Self metaphors. The research has two main goals: the first is to build an Essential Self multilingual WordNet, which serves as the basic lexical resource for Artificial Intelligence describes the core of the human nature. The second is to create a multilingual 3D semantic network.","PeriodicalId":259975,"journal":{"name":"Proceedings of the 6th International Conference on Natural Language Processing and Knowledge Engineering(NLPKE-2010)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-09-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129526111","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Shui nationality characters stroke shape input method 水族文字笔画形状输入法
Hanyue Yang, Xiaorong Chen
{"title":"Shui nationality characters stroke shape input method","authors":"Hanyue Yang, Xiaorong Chen","doi":"10.1109/NLPKE.2010.5587840","DOIUrl":"https://doi.org/10.1109/NLPKE.2010.5587840","url":null,"abstract":"Shape of Shui nationality characters is similar to that of Oracle and Jinwen. In order to work out the problems of how to code hieroglyph, a coding method based on stroke shape for Shui Nationality characters is proposed. The shapes of 467 Shui Nationality characters in the Common Shui Script Dictionary are analyzed, and seven basic strokes are extracted to consist of main Shui characters. Through the statistical comparison, 21 kinds of stroke shape can be got by subdividing the seven basic strokes. A Shui Nationality character is coded by an ordered sequence composed by three strokes taken from the corner of the character according to the coding rules. Finally, the users who can not read the Shui character can input it easily and quickly.","PeriodicalId":259975,"journal":{"name":"Proceedings of the 6th International Conference on Natural Language Processing and Knowledge Engineering(NLPKE-2010)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-09-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130918935","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Chinese patent retrieval based on the pragmatic information 基于语用信息的中文专利检索
Liping Wu, Song Liu, F. Ren
{"title":"Chinese patent retrieval based on the pragmatic information","authors":"Liping Wu, Song Liu, F. Ren","doi":"10.1109/NLPKE.2010.5587776","DOIUrl":"https://doi.org/10.1109/NLPKE.2010.5587776","url":null,"abstract":"In this paper, we propose a novel information retrieval approach based on the pragmatic information for Chinese patents. At present, patent retrieval is becoming more and more important. Not only because patents are always can an important resource in all kinds of field, but patent retrieval save a great deal of time and funds for corporations and researchers. However, with available methods the precision of retrieval results for patents is not very high. What's more, through analyzed the patent documentations we found that except the literal meanings, there are deeper meanings which can be concluded from the patents. Here we call the deeper meanings as pragmatic information. Therefore we established a patent retrieval system to integrate the pragmatic information with classical information retrieval technique to improve the retrieval accuracy. Some experiments using the proposed method have carried out, and the results show that the precision of patent retrieval based on the pragmatic information is higher than the one without using it.","PeriodicalId":259975,"journal":{"name":"Proceedings of the 6th International Conference on Natural Language Processing and Knowledge Engineering(NLPKE-2010)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-09-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125546479","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Part-of-speech tagging for Chinese unknown words in a domain-specific small corpus using morphological and contextual rules 基于形态和语境规则的小语料库中文未知词词性标注
Tao-Hsing Chang, Fu-Yuan Hsu, Chia-Hoang Lee, Hahn-Ming Lee
{"title":"Part-of-speech tagging for Chinese unknown words in a domain-specific small corpus using morphological and contextual rules","authors":"Tao-Hsing Chang, Fu-Yuan Hsu, Chia-Hoang Lee, Hahn-Ming Lee","doi":"10.1109/NLPKE.2010.5587771","DOIUrl":"https://doi.org/10.1109/NLPKE.2010.5587771","url":null,"abstract":"Many studies have tried to search useful information on the Internet by meaningful terms or words. The performance of these approaches is often affected by the accuracy of unknown word extraction and POS tagging, while the accuracy is affected by the size of training corpora and the characteristics of language. This work proposes and develops a method that concentrates on tagging the POS of Chinese unknown words for the domain of our interest, based on the integration of morphological, contextual rules and a statistics-based method. Experimental results indicate that the proposed method can overcome the difficulties resulting from small corpora in oriental languages, and can accurately tags unknown words with POS in domain-specific small corpora.","PeriodicalId":259975,"journal":{"name":"Proceedings of the 6th International Conference on Natural Language Processing and Knowledge Engineering(NLPKE-2010)","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-09-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125283724","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Statistical parsing based on Maximal Noun Phrase pre-processing 基于最大名词短语预处理的统计分析
Qiaoli Zhou, Yue Gu, Xin Liu, Wenjing Lang, Dongfeng Cai
{"title":"Statistical parsing based on Maximal Noun Phrase pre-processing","authors":"Qiaoli Zhou, Yue Gu, Xin Liu, Wenjing Lang, Dongfeng Cai","doi":"10.1109/NLPKE.2010.5587850","DOIUrl":"https://doi.org/10.1109/NLPKE.2010.5587850","url":null,"abstract":"According to the characteristics of Chinese language, this paper proposes a statistical parsing method based on Maximal Noun Phrase(MNP) per-processing. MNP parsing is preferable to be separated from parsing of the full sentence. Firstly, MNP in a sentence are identified; next, MNP can be represented by the head of MNP, and then the sentence is parsed with the head of the MNP. Therefore, the original sentence is divided into two parts, which can be parsed separately. The first part is MNP parsing; the second part is parsing of the sentence in which the MNP are replaced by their head words. Finally, the paper takes Conditional Random Fields (CRFs) as the statistical recognition model of each level in syntactic parsing process.","PeriodicalId":259975,"journal":{"name":"Proceedings of the 6th International Conference on Natural Language Processing and Knowledge Engineering(NLPKE-2010)","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-09-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127018013","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Bagging to find better expansion words 寻找更好的扩展词
Bingqing Wang, Yaqian Zhou, Xipeng Qiu, Qi Zhang, Xuanjing Huang
{"title":"Bagging to find better expansion words","authors":"Bingqing Wang, Yaqian Zhou, Xipeng Qiu, Qi Zhang, Xuanjing Huang","doi":"10.1109/NLPKE.2010.5587826","DOIUrl":"https://doi.org/10.1109/NLPKE.2010.5587826","url":null,"abstract":"The supervised learning has been applied into the query expansion techniques, which trains a model to predict the “goodness” or “utility” of the expanded term to the retrieval system. There are many features to measure the relatedness between the expanded word and the query, which can be incorporated in the supervised learning to select the expanded terms. The training data set is generated automatically by a tricky method. However, this method can be affected by many aspects. A severe problem is that the distribution of the features is query-dependent, which has not been discussed in previous work. With a different distribution on the features, it is questionable to merge these training instances together and use the whole data set to train one single model. In this paper, we first investigate the statistical distribution of the auto-generated training data and show the problems in the training data set. Based on our analysis, we proposed to use the bagging method to ensemble several regression models in order to get a better supervised model to make prediction on the expanded terms. We conducted the experiments on the TREC benchmark test collections. Our analysis on the training data reveals some interesting phenomena about the query expansion techniques. The experiment results also show that the bagging approach can achieve the state-of-art retrieval performance on the standard TREC data set.","PeriodicalId":259975,"journal":{"name":"Proceedings of the 6th International Conference on Natural Language Processing and Knowledge Engineering(NLPKE-2010)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-09-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125257444","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Affix-augmented stem-based language model for persian 波斯语词缀增强词干语言模型
Heshaam Faili, H. Ravanbakhsh
{"title":"Affix-augmented stem-based language model for persian","authors":"Heshaam Faili, H. Ravanbakhsh","doi":"10.1109/NLPKE.2010.5587823","DOIUrl":"https://doi.org/10.1109/NLPKE.2010.5587823","url":null,"abstract":"Language modeling is used in many NLP applications like machine translation, POS tagging, speech recognition and information retrieval. It assigns a probability to a sequence of words. This task becomes a challenging problem for high inflectional languages. In this paper we investigate standard statistical language models on the Persian as an inflectional language. We propose two variations of morphological language models that rely on a morphological analyzer to manipulate the dataset before modeling. Then we discuss shortcoming of these models, and introduce a novel approach that exploits the structure of the language and produces more accurate. Experimental results are encouraging especially when we use n-gram models with small training dataset.","PeriodicalId":259975,"journal":{"name":"Proceedings of the 6th International Conference on Natural Language Processing and Knowledge Engineering(NLPKE-2010)","volume":"54 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-09-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134173801","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Sentiment word identification using the maximum entropy model 基于最大熵模型的情感词识别
Xiaoxu Fei, Huizhen Wang, Jingbo Zhu
{"title":"Sentiment word identification using the maximum entropy model","authors":"Xiaoxu Fei, Huizhen Wang, Jingbo Zhu","doi":"10.1109/NLPKE.2010.5587811","DOIUrl":"https://doi.org/10.1109/NLPKE.2010.5587811","url":null,"abstract":"This paper addresses the issue of sentiment word identification given an opinionated sentence, which is very important in sentiment analysis tasks. The most common way to tackle this problem is to utilize a readily available sentiment lexicon such as HowNet or SentiWordNet to determine whether a word is a sentiment word. However, in practice, words existing in the lexicon sometimes can not express sentiment tendency in a certain context while other words out of the lexicon do express. To address this challenge, this paper presents an approach based on maximum-entropy classification model to identify sentiment words given an opinionated sentence. Experimental results show that our approach outperforms baseline lexicon-based methods.","PeriodicalId":259975,"journal":{"name":"Proceedings of the 6th International Conference on Natural Language Processing and Knowledge Engineering(NLPKE-2010)","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-09-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133917926","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
A reranking method for syntactic parsing with heterogeneous treebanks 异构树库句法分析的重排序方法
Haibo Ding, Muhua Zhu, Jingbo Zhu
{"title":"A reranking method for syntactic parsing with heterogeneous treebanks","authors":"Haibo Ding, Muhua Zhu, Jingbo Zhu","doi":"10.1109/NLPKE.2010.5587842","DOIUrl":"https://doi.org/10.1109/NLPKE.2010.5587842","url":null,"abstract":"In the field of natural language processing (NLP), there often exist multiple corpora with different annotation standards for the same task. In this paper, we take syntactic parsing as a case study and propose a reranking method which is able to make direct use of disparate treebanks simultaneously without using techniques such as treebank conversion. The method proceeds in three steps: 1) build parsers on individual treebanks; 2) use parsers independently to generate n-best lists for each sentence in test set; 3) rerank individual n-best lists which correspond to the same sentence by using consensus information exchanged among these n-best lists. Experimental results on two open Chinese treebanks show that our method significantly outperforms the baseline system by 0.84% and 0.53% respectively.","PeriodicalId":259975,"journal":{"name":"Proceedings of the 6th International Conference on Natural Language Processing and Knowledge Engineering(NLPKE-2010)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-09-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123574465","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Flexible English writing support based on negative-positive conversion method 基于正负转换法的灵活英语写作支持
Yasushi Katsura, Kazuyuki Matsumoto, F. Ren
{"title":"Flexible English writing support based on negative-positive conversion method","authors":"Yasushi Katsura, Kazuyuki Matsumoto, F. Ren","doi":"10.1109/NLPKE.2010.5587778","DOIUrl":"https://doi.org/10.1109/NLPKE.2010.5587778","url":null,"abstract":"With development of the recent globalization, the chance to exchange in English increased in the business field. In particular, it's necessary to write a thesis and a charter handwriting in English. Because many Japanese are not used to making English sentence, it is a great burden to write appropriate sentence in English without any support for creating English sentence. In this study we have developed an English composition support system. By this system, it's to search for the interlinear translation example to refer to by database and generate a new sentence by replacing a noun in the example sentence. In this paper, based on the technique of Super-Function, we propose a method to convert an affirmative sentence into negative sentence and vice versa to realize more flexible and extensive text conversion.","PeriodicalId":259975,"journal":{"name":"Proceedings of the 6th International Conference on Natural Language Processing and Knowledge Engineering(NLPKE-2010)","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-09-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121698126","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信