Fausto Fleites, S. Cocke, Shu‐Ching Chen, S. Hamid
{"title":"Efficiently integrating MapReduce-based computing into a Hurricane Loss Projection model","authors":"Fausto Fleites, S. Cocke, Shu‐Ching Chen, S. Hamid","doi":"10.1109/IRI.2013.6642499","DOIUrl":"https://doi.org/10.1109/IRI.2013.6642499","url":null,"abstract":"Homeowner insurance is a critical issue for Floridians because of the periodic threat hurricanes pose to Florida. Providing fairness into the rate-making policy process, the state of Florida has developed the Florida Public Hurricane Loss Model (FPHLM), an open, public hurricane risk model to assess the risk of wind damage to insured residential properties. For each input property portfolio, the FPHLM processes a large amount of data to provide expected losses over tens of thousand of years of simulation, for which computational efficiency is of paramount importance. This paper presents our work in integrating the atmospheric component into the FPHLM using MapReduce, which resulted in a highly efficient computing platform for generating stochastic hurricane events on a cluster of computers. The experimental results demonstrate the feasibility of utilizing MapReduce for risk modeling components.","PeriodicalId":418492,"journal":{"name":"2013 IEEE 14th International Conference on Information Reuse & Integration (IRI)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-10-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132098102","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A knowledge-based semantic tool for standard part management in aerospace industry","authors":"Hasan Mert Taymaz, Tansel Özyer, C. Cangelir","doi":"10.1109/IRI.2013.6642528","DOIUrl":"https://doi.org/10.1109/IRI.2013.6642528","url":null,"abstract":"Manufacture-based high-tech industries such as automotive and aerospace essentially pay attention to reducing costs and improving the quality of their products. In this context, assembly of products is one prominent part of finalizing the product. During the assembly, finding the most suitable standard parts1 that combines one or more bodies of assembly to form the final product is the common task for design engineers. Quite a few relational standard part databases can be found on the market, however none of them properly address the design needs. While querying relational databases, results are mostly vague or irrelevant, hence it requires domain experience about the features of parts. On the other hand semantic search [1] by the help of graphic model, it is possible to navigate and access all standard parts on graph; it returns relevant results as usual. In this paper, we propose semantic based search engine for inquiring standard parts from the repository. In our approach, we model and store Standard Part data on a commercial RDF (Resource Description Framework) Database [2]. We used OWLIM [3] to store data on database and we created an intelligent search engine, which discover link relations on RDF data. Our prototype system is still being developed in collaboration with the TOBB University. Final product will be used in Turkish Aerospace Industries (TAI)2 to resolve their critical problems.","PeriodicalId":418492,"journal":{"name":"2013 IEEE 14th International Conference on Information Reuse & Integration (IRI)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-10-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131045143","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"An approach to develop frameworks from feature models","authors":"M. C. Viana, R. Penteado, A. F. Prado, R. Durelli","doi":"10.1109/IRI.2013.6642523","DOIUrl":"https://doi.org/10.1109/IRI.2013.6642523","url":null,"abstract":"Frameworks are reusable software composed of concrete and abstract classes that implement the functionality of a domain. Applications can reuse framework design and code in order to improve their quality and be developed more efficiently. However, to develop software for reuse, such as a framework, is harder than to develop an application. Hence, in this paper we present an approach, named From Features to Framework (F3), to facilitate the development of white box frameworks. This approach is divided in two steps: Domain Modeling, in which the features of the framework are defined; and Framework Construction, in which the framework is designed and implemented according to its features and their relationships. We also present an experiment that evaluated the F3 approach showing that it makes framework development easier and more efficient.","PeriodicalId":418492,"journal":{"name":"2013 IEEE 14th International Conference on Information Reuse & Integration (IRI)","volume":"57 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-10-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131300818","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Patient response datasets: Challenges and opportunities","authors":"Randall Wald, T. Khoshgoftaar","doi":"10.1109/IRI.2013.6642480","DOIUrl":"https://doi.org/10.1109/IRI.2013.6642480","url":null,"abstract":"As the field of bioinformatics has grown in importance, more and more studies have investigated the use of gene microarray datasets to understand cancer. Although much of this research has focused on which genes are differently-expressed between cancerous and non-cancerous tissues, an equally important question is which genes are most useful for predicting the success of cancer treatment. How well a patient will respond to a given treatment depends on the specifics of their cancer, and biopsies alone cannot detect genetic markers; thus, gene chips are an increasingly valuable research tool in this field. The problem of identifying which gene markers are predictive of successful response to treatment differs from more general cancer-identification and cancer-classification problems due to the expected similarities among all patients (since they share a cancer type), and thus it is important to understand this collection of datasets as a group separate from other cancer-related microarray datasets. The present work surveys research using gene microarray datasets for the task of patient response prediction, in particular research which uses this data to discover the best genes for predicting patient response. We discuss the methods and procedures of the surveyed papers and which approaches were used for gene selection, and present ideas and strategies for future work which further explores how to best identify gene signatures which can be used clinically to help select the best treatment for a given cancer.","PeriodicalId":418492,"journal":{"name":"2013 IEEE 14th International Conference on Information Reuse & Integration (IRI)","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-10-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114612294","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Chinese textual entailment with Wordnet semantic and dependency syntactic analysis","authors":"Chun-yung Tu, Min-Yuh Day","doi":"10.1109/IRI.2013.6642455","DOIUrl":"https://doi.org/10.1109/IRI.2013.6642455","url":null,"abstract":"Recognizing Inference in TExt (RITE) is a task for automatically detecting entailment, paraphrase, and contradiction in texts which addressing major text understanding in information access research areas. In this paper, we proposed a Chinese textual entailment system using Wordnet semantic and dependency syntactic approaches in Recognizing Inference in Text (RITE) using the NTCIR-10 RITE-2 subtask datasets. Wordnet is used to recognize entailment at lexical level. Dependency syntactic approach is a tree edit distance algorithm applied on the dependency trees of both the text and the hypothesis. We thoroughly evaluate our approach using NTCIR-10 RITE-2 subtask datasets. As a result, our system achieved 73.28% on Traditional Chinese Binary-Class (BC) subtask and 74.57% on Simplified Chinese Binary-Class subtask with NTCIR-10 RITE-2 development datasets. Thorough experiments with the text fragments provided by the NTCIR-10 RITE-2 subtask showed that the proposed approach can improve system's overall accuracy.","PeriodicalId":418492,"journal":{"name":"2013 IEEE 14th International Conference on Information Reuse & Integration (IRI)","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-10-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114694210","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Detecting data exfiltration by integrating information across layers","authors":"Puneet Sharma, A. Joshi, Timothy W. Finin","doi":"10.1109/IRI.2013.6642487","DOIUrl":"https://doi.org/10.1109/IRI.2013.6642487","url":null,"abstract":"Data exfiltration is the unauthorized leakage of confidential data from a system. Unlike intrusions that seek to overtly disable or damage a system, it is particularly hard to detect because it uses a variety of low/slow vectors and advanced persistent threats (APTs). It is often assisted (intentionally or not) by an insider who might be an employee who downloads a trojan or uses a hardware component that has been tampered with or acquired from an unreliable source. Conventional scan and test based detection approaches work poorly, especially for hardware with embedded trojans. We describe a framework to detect potential exfiltration events that actively monitors of a set of key parameters that cover the entire stack, from hardware to the application layer. An attack alert is generated only if several monitors detect suspicious activity within a short temporal window. The cross-layer monitoring and integration helps ensure accurate alerts with fewer false positives and makes designing a successful attack more difficult.","PeriodicalId":418492,"journal":{"name":"2013 IEEE 14th International Conference on Information Reuse & Integration (IRI)","volume":"55 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-10-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116654346","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Filter- and wrapper-based feature selection for predicting user interaction with Twitter bots","authors":"Randall Wald, T. Khoshgoftaar, Amri Napolitano","doi":"10.1109/IRI.2013.6642501","DOIUrl":"https://doi.org/10.1109/IRI.2013.6642501","url":null,"abstract":"High dimensionality (the presence of too many features) is a problem which plagues many datasets, including mining from personality profiles. Feature selection can be used to reduce the number of features, and many strategies have been proposed to help select the most important features from a larger group. Feature rankers will produce a metric for each feature and return the best for a given subset size, while filter-based subset evaluation will perform statistical analysis on whole subsets and wrapper-based subset selection will use classification models with chosen features to decide which are most important for model-building. While all three approaches have been discussed in the literature, relatively little work compares all three with one another directly. In the present study, we do precisely this, considering feature ranking, filter-based subset evaluation, and wrapper-based subset selection (along with no feature ranking) on two datasets based on predicting interaction with bots on Twitter. For the two subset-based techniques, we consider two search techniques (Best First and Greedy Stepwise) to build the subsets, while we use one feature ranker (ROC) chosen for its excellent performance in previous works. Six learners are used to build models with the selected features. We find that feature ranking consistently performs well, giving the best results for four of the six learners on both datasets. In addition, all of the techniques other than feature ranking perform worse than no feature selection for four of six learners. This leads us to recommend the use of feature ranking over more complex subset evaluation techniques.","PeriodicalId":418492,"journal":{"name":"2013 IEEE 14th International Conference on Information Reuse & Integration (IRI)","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-10-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127757222","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Federated ensemble Kalman filter in no reset mode design","authors":"M. Kazerooni, F. Shabaninia, M. Vaziri, S. Vadhva","doi":"10.1109/IRI.2013.6642539","DOIUrl":"https://doi.org/10.1109/IRI.2013.6642539","url":null,"abstract":"The main contribution of this paper is to design a more accurate optimal/suboptimal fault tolerant state estimator. Federated filters compose of a set of local filters and a master filter, the local filters work in parallel and their solutions are periodically fused by the master filter yielding a global solution. Federated ensemble Kalman filter no reset configuration is developed for multi-sensor data fusion. Ensemble Kalman filter(ENKF) estimation is widely used, where the models are of extremely high order and nonlinear, the initial states are highly uncertain, and a large number of measurements are available. ENKF is used as local filters in federated filter no reset mode design. Fault detection and isolation (FDI) algorithms is applied to local filter's outputs. Faulty local filters are isolated and not fused by master filter to get a fault tolerant filter. Simulation results demonstrate the validity of the proposed filter formation.","PeriodicalId":418492,"journal":{"name":"2013 IEEE 14th International Conference on Information Reuse & Integration (IRI)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-10-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129359699","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
M. Saleem, Waqas Nawaz, Young-Koo Lee, Sungyoung Lee
{"title":"Road segment partitioning towards anomalous trajectory detection for surveillance applications","authors":"M. Saleem, Waqas Nawaz, Young-Koo Lee, Sungyoung Lee","doi":"10.1109/IRI.2013.6642525","DOIUrl":"https://doi.org/10.1109/IRI.2013.6642525","url":null,"abstract":"Recently, the low cost and high availability of location acquisition technologies has significantly increased the demands for online anomalous trajectory detection. It is being used in social as well as commercial areas to provide human life care applications like healthcare, theft protection and taxi fraud detection. However, anomalous trajectory detection is still a challenging problem. The main complications involved in it, are inaccuracy in obtaining trajectory traces and evaluation of partial anomalous trajectories. In this study we contribute towards resolving these complications by proposing a novel method of Road segment Partitioning towards Anomalous Trajectory Detection (RPat). Our proposed method partitions the trajectory on the basis of road segments. Then, these sub-trajectories are evaluated, independently based on contemporary behavior of moving objects to accurately analyze the trajectories that possess abnormal behavior at any intermediate parts. The evaluation score of each sub-trajectory is aggregated to reflect the attitude of an overall itinerary as anomalous or regular. Further, the accuracy in the reconstruction of trajectories is achieved by plotting the itinerary traces on real world road maps. Experimental studies are conducted on real datasets and an accuracy of more than 81% is achieved.","PeriodicalId":418492,"journal":{"name":"2013 IEEE 14th International Conference on Information Reuse & Integration (IRI)","volume":"107 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-10-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132442856","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Natural image segmentation using morphological mathematics and fuzzy logic","authors":"Victoria L. Fox, M. Milanova","doi":"10.1109/IRI.2013.6642542","DOIUrl":"https://doi.org/10.1109/IRI.2013.6642542","url":null,"abstract":"The segmentation of natural images remains a challenging task in image processing. Many methods have been proposed in the literature regarding algorithms for the segmentation of such images. Many of the algorithms are complex in nature and inefficient in practice with unaltered images. In order to efficiently use the algorithms it is beneficial to preprocess the natural images. However, natural images often involve subjects and background that are not easily quantified with crisp preprocessing parameters. To this, we will show the use of grey-scale morphological operators coupled with fuzzy image enhancement with natural images is an efficient and noncomplex method that more accurately isolates the region of interest in the image and will define a novel combination of grey-scale morphological operators for use with natural images.","PeriodicalId":418492,"journal":{"name":"2013 IEEE 14th International Conference on Information Reuse & Integration (IRI)","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-10-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115421819","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}