{"title":"A study on e-mail image spam filtering techniques","authors":"S. Dhanaraj, V. Karthikeyani","doi":"10.1109/ICPRIME.2013.6496446","DOIUrl":"https://doi.org/10.1109/ICPRIME.2013.6496446","url":null,"abstract":"Spam filters are the most used software tool used by businesses and individual to block spam mails entering into their mail boxes. Until recently, majority of research effort was expended on controlling text-based spam emails. However, the past few years have envisaged a novel approach, where the spammers embed the text message into an image. Thus, the anti-spam filtering research is forced to move from text-based techniques to image-based techniques. Spam and the spam blockers designed to combat it have spawned an upsurge in creativity and innovation. Many software developers are developing new and every more effective spam filtering software. All the methods have a common dream that is to eliminate 100% of the spam, which is still not a reality. To reduce the gap between this reality and dream, researchers have proposed many different types of spam filters and this paper provides a review of them.","PeriodicalId":123210,"journal":{"name":"2013 International Conference on Pattern Recognition, Informatics and Mobile Engineering","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-04-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126554189","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Object segmentation through optimized algorithm","authors":"S. Thilagamani, N. Shanthi","doi":"10.1109/ICPRIME.2013.6496438","DOIUrl":"https://doi.org/10.1109/ICPRIME.2013.6496438","url":null,"abstract":"This paper gives a clear picture about segmenting the object from an image through the process of pattern matching. It is challenge to retrieve desired object from an image. The various steps involved are preprocessing the input image, forming the super pixel and segmenting the desired object.","PeriodicalId":123210,"journal":{"name":"2013 International Conference on Pattern Recognition, Informatics and Mobile Engineering","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-04-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128741937","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Data property analyzer for information storage in cloud","authors":"S. Srinivasan, R. Krishnan","doi":"10.1109/ICPRIME.2013.6496518","DOIUrl":"https://doi.org/10.1109/ICPRIME.2013.6496518","url":null,"abstract":"This paper inscribes a new approach for maintaining the data usage report for the cloud data storage using a novel data property analyzer. Cloud data storage is a technology that uses the internet and central remote servers to maintain data and share the applications. It allows consumer to use applications without installation and access their personal files at any computer with internet access. In general data property analysis system, Source and destination file content is compared in the form of bytes. In the cloud environment, data verification is needed for every computation in the storage correctness. So every time the data is retrieved from local system and compared with the destination file from the cloud zone. This procedure takes too much of time to find out a tiny change in the cloud file content. The proposed system is implemented with the idea of checking the file properties to find out the change in file content instead of verifying the entire file content. To check properties we have taken some of the file attributes such as file size, file modification date, file name and location.","PeriodicalId":123210,"journal":{"name":"2013 International Conference on Pattern Recognition, Informatics and Mobile Engineering","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-04-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122223411","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Contour optimization of suspension insulators using real coded genetic algorithm with simulated binary crossover","authors":"S. Gunasekaran, M. Iruthayarajan","doi":"10.1109/ICPRIME.2013.6496501","DOIUrl":"https://doi.org/10.1109/ICPRIME.2013.6496501","url":null,"abstract":"Electrical-field distribution along the insulator surface strongly depends upon the contour design, besides the effect of pollution. The insulator contour should be designed to reach a desired uniform and minimal tangential field to increase the onset voltage of surface flashover. In this paper, with the finite element method (FEM) integrated, the real coded genetic algorithm (RGA) with simulated binary crossover (SBX) approach is proposed for contour optimization of a suspension insulator. The aim of the contour optimization is to minimize the Tangential electric field and make the tangential electric field as uniform, subject to design constraints. The results show that a rather uniform and minimal tangential field distribution can be obtained through the proposed approach.","PeriodicalId":123210,"journal":{"name":"2013 International Conference on Pattern Recognition, Informatics and Mobile Engineering","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-04-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130657457","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"An adaptive assessment system to compose serial test sheets using item response theory","authors":"Kavitha Rajamani, V. Kathiravan","doi":"10.1109/ICPRIME.2013.6496458","DOIUrl":"https://doi.org/10.1109/ICPRIME.2013.6496458","url":null,"abstract":"The conventional learning environment is being rapidly supplemented by an E-Learning environment, particularly Computer Assisted Instruction (CAI). Each learner has different learning status and therefore should use different test items in their evaluation. The Computerized Adaptive Test (CAT) can adjust the degree of difficulty of test items dynamically depending on their ability. One of the most important and challenging issues in conducting a good test is the construction of test sheets that can meet various criteria. Therefore, several measures have been proposed here to represent the quality of each test item. In E-Learning, there is a critical need of formative assessment and hence series of test sheets to be composed which pose some challenges. Consequently the selection of appropriate test items is important when constructing an assessment sheet that meets multi-criteria assessment requirements, such as expected difficulty degree, expected discrimination degree, specified distribution of relevant concept weights and exposure frequency of items. All behavior like selection of items, evaluation and termination of test are based on Item Response Theory, an educational theory. An adaptive assessment system is proposed to compose serial test sheets. This approach will improve the efficiency of composing near optimal serial test items to meet multiple assessment criteria.","PeriodicalId":123210,"journal":{"name":"2013 International Conference on Pattern Recognition, Informatics and Mobile Engineering","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-04-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117248934","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A novel approach for speech feature extraction by Cubic-Log compression in MFCC","authors":"M. R. Devi, T. Ravichandran","doi":"10.1109/ICPRIME.2013.6496469","DOIUrl":"https://doi.org/10.1109/ICPRIME.2013.6496469","url":null,"abstract":"Speech Pre-processing is measured as major step in development of feature vector extraction for an efficient Automatic Speech Recognition (ASR) system. A novel approach for speech feature extraction is by applying the Mel-frequency cepstral co-efficient (MFCC) algorithm using Cubic-Log compression instead of Logarithmic compression in MFCC. In proposed MFCC, the frequency axis is initially warped to the mel-scale which is roughly below 2 kHz and logarithmic above this point. Triangular filter are equally spaced in the mel-scale are applied on the warped spectrum. The result of the filters are compressed using Cubic-Log function and cepstral co-efficient are computed by applying DCT to obtain minimum MFCC feature vector for spoken words. These feature vectors are given as input to classification and Recognition phase. The system is trained and tested by generating MFCC feature vector for 600 isolated words, 256 connected words and 150 sentences in clear and noisy environment. Experiment results shows that with minimum MFCC feature vector is enough for speech recognition system to achieve high recognition rate and its performance is measured based on Mean Square Error (MSE) rate.","PeriodicalId":123210,"journal":{"name":"2013 International Conference on Pattern Recognition, Informatics and Mobile Engineering","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-04-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133238584","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Dual transform based steganography using wavelet families and statistical methods","authors":"G. Prabakaran, R. Bhavani, K. Kanimozhi","doi":"10.1109/ICPRIME.2013.6496488","DOIUrl":"https://doi.org/10.1109/ICPRIME.2013.6496488","url":null,"abstract":"Steganography is the discipline of exchanging top secret information by embedding it into a multimedia carrier. The ultimate aim, here is to hide the very existence of the embedded information within seemingly innocuous carriers. The proposed method extracts either Discrete Wavelet Transform (DWT) or Integer Wavelet Transform (IWT) coefficients of both cover image and secret image. After that two extracted coefficient values are embedded by fusion processing technique. Then the stego image is obtained by applying various combinations of DWT and IWT on both images. In this method, we concentrated for perfecting the visual effect of the stego image and robustness against the various attacks by using different wavelet families. Finally performance evaluation is done on dual transform steganography using wavelet families and statistical methods. In our method achieved acceptable imperceptibility and certain robustness.","PeriodicalId":123210,"journal":{"name":"2013 International Conference on Pattern Recognition, Informatics and Mobile Engineering","volume":"51-52 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-04-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123011773","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A review and design study of cross layer scheme based algorithm to reduce the link break in MANETs","authors":"R. Kumar, P. Kamalakkannan","doi":"10.1109/ICPRIME.2013.6496462","DOIUrl":"https://doi.org/10.1109/ICPRIME.2013.6496462","url":null,"abstract":"Mobile an ad hoc network is a self organizing wireless network for built spontaneously to connect the mobile devices. It does not require any fixed infrastructure due to no wired backbone. It is suitable to use in environment that have need of on the fly set-up. Every host is a router and packet forwarder. Each node may be mobile, and topology changes frequently and unpredictably due to the arbitrary mobility of mobile nodes. This aspect leads to frequent path failure and route rebuilding. Routing protocol development depends on mobility management, efficient bandwidth and power usage which are critical in ad hoc networks. Cross layer scheme is the new method for wireless mobile devices become to increase the capacity network. In this paper, proposed a new cross layer scheme based algorithm to reduce the link break in MANETs. This present three schemes to reduce packet retransmission ratio by distribution signal information between PHYSICAL LAYER and MAC LAYER, and to discuss the frequent route failures in MANET by identify shortly to be broken links instead of prediction of received signal power, To find optimized route maintenance by taking into consideration of Bandwidth, Delay which results by improvement of QoS. The performance of cross layer process for all the schemes simulation are carried out to evaluate network performance using network simulator NS-2 based on the basic parameters like Throughput, packet delivery ratio(PDR) in term of number of nodes when various pause times.","PeriodicalId":123210,"journal":{"name":"2013 International Conference on Pattern Recognition, Informatics and Mobile Engineering","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-04-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115225618","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"On modular multiplicative divisor graphs","authors":"R. Revathi, R. Rajeswari","doi":"10.1109/ICPRIME.2013.6496447","DOIUrl":"https://doi.org/10.1109/ICPRIME.2013.6496447","url":null,"abstract":"A graph G (V, E) with /V/ = n is said to have Modular Multiplicative Divisor (MMD) labeling if there exist a bijection f: V(G)→{1, 2, n} and the induced function f*: E(G)→{0, 1, 2, ..., n-1} where f*(uv) = f(u)f(v) mod n such that n divides the sum of all edge labels of G. In this paper we prove that the split graph of cycle C<sub>n</sub>, helm graph H<sub>n</sub>, flower graph f<sub>nX4</sub>, cycle cactus C<sub>4</sub> (n) and extended triplicate graph of a path P<sub>n</sub> (ETG(P<sub>n</sub>)) admits Modular Multiplicative Divisor labeling. AMS Subject Classification: 05C78","PeriodicalId":123210,"journal":{"name":"2013 International Conference on Pattern Recognition, Informatics and Mobile Engineering","volume":"179 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-04-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115225882","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A fault tolerent approach in scientific workflow systems based on cloud computing","authors":"K. Ganga, S. Karthik","doi":"10.1109/ICPRIME.2013.6496507","DOIUrl":"https://doi.org/10.1109/ICPRIME.2013.6496507","url":null,"abstract":"Fault Tolerance is a configuration that prevents a computer or network device from failing in the event of unexpected problem or error such as hardware failure, link failure, unauthorized access, variations in the configuration of different systems and system running out of memory or disk space. The integration of fault tolerance measures with scheduling gains much importance. Scientific workflows use distributed heterogeneous resources in cloud interface are often hard to program. This paper explains task replication technique and simulation of cloud computing systems.","PeriodicalId":123210,"journal":{"name":"2013 International Conference on Pattern Recognition, Informatics and Mobile Engineering","volume":"184 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-04-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114753173","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}