{"title":"Building a Machine Translation System in a Restrict Context from Ka-Tu Language into Vietnamese","authors":"H. Le, Phan Thi Bong, H. Phan","doi":"10.1109/KSE.2012.26","DOIUrl":"https://doi.org/10.1109/KSE.2012.26","url":null,"abstract":"The paper introduces some basic characteristics of Ka Tu language, from that we built the machine translation system from Ka Tu language into Vietnamese in a restrict context for translating the weather forecast bulletins, supporting the communication of Ka Tu language of Vietnamese voice emission. The system was tested at the Quang Nam Radio and Television station. The initial Results are Positive for the test of the system. The machine translation system from Ka Tu language into Vietnamese in a restrict context contributed to solving the serious lack of information in aboriginal languages and can be expanded into other areas such as warming fire alarm, broadcasting policies and laws of the State, farming experience, animal husbandry, natural disaster prevetion... Because the corpus with open structure, the optical axis is in Vietnamese language, a multilingual corpus can be easily built by adding other ethnic languages such as Cham, Ede, Jarai, Muong,...","PeriodicalId":122680,"journal":{"name":"2012 Fourth International Conference on Knowledge and Systems Engineering","volume":"133 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-08-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115635057","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A Multi-point Interactive Method for Multi-objective Evolutionary Algorithms","authors":"Long Nguyen, L. Bui","doi":"10.1109/KSE.2012.30","DOIUrl":"https://doi.org/10.1109/KSE.2012.30","url":null,"abstract":"Many real-world optimization problems have more than one objective (and these objectives are often conflicting). In most cases, there is no single solution being optimized with regards to all objectives. Deal with such problems, Multi-Objective Evolutionary Algorithms (MOEAs) have shown a great potential. There has been a popular trend in getting suitable solutions and increasing the convergence of MOEAs, that is consideration of Decision Maker (DM) during the optimization process (interacting with DM) for checking, analyzing the results and giving the preference. In this paper, we propose an interactive method allowing DM to specify a set of reference points. It used a generic algorithm framework of MOEA/D, a widely-used and decomposition-based MOEA for demonstration of concept. Basically MOEA/D decomposes a multi-objective optimization problem into a number of different single-objective optimization sub-problems and defines neighborhood relations among these sub-problems. Then a population-based method is used to optimize these sub-problems simultaneously. Each sub-problem is optimized by using information mainly from its neighboring sub-problems. In MOEA/D an ideal point is used to choose neighbored solutions for each run. Instead of using a single point, we introduce an alternative to the set of reference points. There are several way to take into account the information of the region specified by the set of reference points; here we used the mean of this set (or we call the combined point). The combined point which represents for the set of reference points from DM is used either to replace or adjust the current ideal point obtained by MOEA/D. We carried out a case study on several test problems and obtained quite good results.","PeriodicalId":122680,"journal":{"name":"2012 Fourth International Conference on Knowledge and Systems Engineering","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-08-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123785386","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A Study on Prosody of Vietnamese Emotional Speech","authors":"Thi Duyen Ngo, T. D. Bui","doi":"10.1109/KSE.2012.19","DOIUrl":"https://doi.org/10.1109/KSE.2012.19","url":null,"abstract":"This paper describes the analyses of the prosody of Vietnamese emotional speech, accomplished to find the relations between prosodic variations and emotional states in Vietnamese speech. These relations were obtained by investigating the variations of prosodic features in Vietnamese emotional speech in comparison with prosodic features of neutral speech. The analyses were performed on a multi-style emotional speech database which consisted of Vietnamese sentences uttered in different styles. Specifically, four emotional styles were considered: happiness, sadness, cold anger, and hot anger. Speech data in the neutral style were also collected, and prosodic differences of each style with respect to this neutral baseline were quantified. The acoustic features related to prosody which were investigated were fundamental frequency, power, and duration. According to the analysis results, for each speaker of the database, a set of prosodic variation coefficients was produced for each emotional style. This will help for bringing emotions into Vietnamese synthesized speech, making them more natural.","PeriodicalId":122680,"journal":{"name":"2012 Fourth International Conference on Knowledge and Systems Engineering","volume":"456 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-08-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123875936","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A Method of Object Identification Based on Fuzzy Object Functional Dependencies in Fuzzy Object-Oriented Databases","authors":"Vuong Quang, Doan Van Ban, H. Ha","doi":"10.1109/KSE.2012.14","DOIUrl":"https://doi.org/10.1109/KSE.2012.14","url":null,"abstract":"In this article, we propose a method of object identification in the state of fuzzy object-oriented database schema based on different notions of fuzzy object functional dependencies. Here, fuzzy object functional dependencies allow to express constraints on attributes of arbitrary object types including the types themselves. Moreover, the different forms of object identification are also introduced and discussed.","PeriodicalId":122680,"journal":{"name":"2012 Fourth International Conference on Knowledge and Systems Engineering","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-08-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128379956","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Thanh-Luong Tran, Quang-Thuy Ha, Thi-Lan-Giao Hoang, Linh Anh Nguyen, H. Nguyen, A. Szałas
{"title":"Concept Learning for Description Logic-Based Information Systems","authors":"Thanh-Luong Tran, Quang-Thuy Ha, Thi-Lan-Giao Hoang, Linh Anh Nguyen, H. Nguyen, A. Szałas","doi":"10.1109/KSE.2012.23","DOIUrl":"https://doi.org/10.1109/KSE.2012.23","url":null,"abstract":"The work [1] by Nguyen and Szalas is a pioneering one that uses bisimulation for machine learning in the context of description logics. In this paper we generalize and extend their concept learning method [1] for description logic-based information systems. We take attributes as basic elements of the language. Each attribute may be discrete or numeric. A Boolean attribute is treated as a concept name. This approach is more general and much more suitable for practical information systems based on description logic than the one of [1]. As further extensions we allow also data roles and the concept constructors \"functionality\" and \"unquantified number restrictions\". We formulate and prove an important theorem on basic selectors. We also provide new examples to illustrate our approach.","PeriodicalId":122680,"journal":{"name":"2012 Fourth International Conference on Knowledge and Systems Engineering","volume":"45 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-08-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130062027","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Sentence Splitting for Vietnamese-English Machine Translation","authors":"Bui Thanh Hung, Minh Le Nguyen, Akira Shimazu","doi":"10.1109/KSE.2012.28","DOIUrl":"https://doi.org/10.1109/KSE.2012.28","url":null,"abstract":"Translation quality is often disappointed when a phrase based machine translation system deals with long sentences. Because of syntactic structure discrepancy between two languages, the translation output will not preserve the same word order as the source. When a sentence is long, it should be partitioned into several clauses and the word reordering in the translation should be done within clauses, not between clauses. In this paper, a rule-based technique is proposed to split long Vietnamese sentences based on linguistic information. We use splitting boundaries for translating sentences with two type of constrains: wall and zone. This method is useful for preserving word order and improving translation quality. We describe experiments on translation from Vietnamese to English, showing an improvement BLEU and NIST score.","PeriodicalId":122680,"journal":{"name":"2012 Fourth International Conference on Knowledge and Systems Engineering","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-08-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130840171","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A Method for Generating Models of Black-Box Components","authors":"Le Ba Cuong, Pham Ngoc Hung","doi":"10.1109/KSE.2012.15","DOIUrl":"https://doi.org/10.1109/KSE.2012.15","url":null,"abstract":"The model-based approaches are difficult to be applied in practice due to the lack of formal models describing behaviors of systems. This paper proposes a method for generating accurate models of components in order to solve this issue in the context of the component-based systems. The key idea of this method is to generate minimal deterministic finite automata as the accurate models of the corresponding components. For this purpose, the proposed method first computes a set of traces as a regular language of a given component by executing all possible experiments over the alphabet of the component. This method then constructs a regular expression to represent this set. After that, a minimal deterministic finite automaton as an accurate model of the component is generated by applying the Thompson algorithm and some optimized activities. The generated models are useful for the existing model-based approaches, e.g., model checking and model-based testing in improving quality of component-based software. An implemented tool supporting the method and experimental results are also presented.","PeriodicalId":122680,"journal":{"name":"2012 Fourth International Conference on Knowledge and Systems Engineering","volume":"202 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-08-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131478798","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A Maximum Likelihood Method for Detecting Bad Samples from Illumina BeadChips Data","authors":"Ha Nguyen, L. Vinh, S. Le","doi":"10.1109/KSE.2012.24","DOIUrl":"https://doi.org/10.1109/KSE.2012.24","url":null,"abstract":"Genotype data provide crucial information to understand effects of genetic variation to human health. Current microarray technologies are able to generate raw genotype data from thousands of samples across million of SNP sites. These raw data are processed by computational methods, called genotype caller, to obtain genotypes. Genotype calls of different callers might not be consistent due to noise of bad samples or SNPs. This requires a manual quality control step conducted by experts to remove bad samples or bad SNP sites. In this paper, we propose a maximum likelihood method to detect bad samples to improve the reliability of the results. Experiments with real data demonstrate the usefulness of our method in the quality control process. Thus, our method has the ability to reduce the number of samples that are requested to manually check by experts.","PeriodicalId":122680,"journal":{"name":"2012 Fourth International Conference on Knowledge and Systems Engineering","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-08-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125395914","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Upscaling and Assessing Information of Agriculture Indicators in Agent-Based Assessment Model from Field to Region Scale","authors":"V. N. Nguyen, A. Drogoul, H. Huynh","doi":"10.1109/KSE.2012.9","DOIUrl":"https://doi.org/10.1109/KSE.2012.9","url":null,"abstract":"Problems related to brown plant hopper outbreaks management in the Mekong Delta-Vietnam are complex and need aggregating information across administrative scales. Brown plant hopper density and environmental indicators are main factors that support stakeholders have suitable decision making. Due to detail levels of information are different from field scale to region scale, upscaling information is required, this paper aims at designing an agent-based assessment model and proposes an upscaling method to aggregate and assess information from field to region scales. A case study presents conveying rice area infested by brown plant hoppers across scales and results could help stakeholders assess damage levels caused by Brown plant hoppers in rice crop seasons.","PeriodicalId":122680,"journal":{"name":"2012 Fourth International Conference on Knowledge and Systems Engineering","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-08-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126506968","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Detecting Bad SNPs from Illumina BeadChips Using Jeffreys Distance","authors":"S. H. Nguyen, L. Vinh, S. Le","doi":"10.1109/KSE.2012.25","DOIUrl":"https://doi.org/10.1109/KSE.2012.25","url":null,"abstract":"Current microarray technologies are able to assay thousands of samples over million of SNPs simultaneously. Computational approaches have been developed to analyse a huge amount of data from microarray chips to understand sophisticated human genomes. The data from microarray chips might contain errors due to bad samples or bad SNPs. In this paper, we propose a method to detect bad SNPs from the probe intensities data of Illumina Beadchips. This approach measures the difference among results determined by three software Illuminus, GenoSNP and Gencall to detect the unstable SNPs. Experiment with SNP data in chromosome 20 of Kenyan people demonstrates the usefulness of our method. This approach reduces the number of SNPs that are needed to check manually. Furthermore, it has the ability in detecting bad SNPs that have not been recognized by other criteria.","PeriodicalId":122680,"journal":{"name":"2012 Fourth International Conference on Knowledge and Systems Engineering","volume":"12 8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-08-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129767276","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}