地理信息学中的重复检测:从区间和模糊数到一般多维不确定性

S. Starks, L. Longpré, R. Araiza, V. Kreinovich, Hung T. Nguyen
{"title":"地理信息学中的重复检测:从区间和模糊数到一般多维不确定性","authors":"S. Starks, L. Longpré, R. Araiza, V. Kreinovich, Hung T. Nguyen","doi":"10.1109/NAFIPS.2007.383900","DOIUrl":null,"url":null,"abstract":"Geospatial databases generally consist of measurements related to points (or pixels in the case of raster data), lines, and polygons. In recent years, the size and complexity of these databases have increased significantly and they often contain duplicate records, i.e., two or more close records representing the same measurement result. In this paper, we address the problem of detecting duplicates in a database consisting of point measurements. As a test case, we use a database of measurements of anomalies in the Earth's gravity field that we have compiled. In our previous papers (2003,2004), we have proposed a new fast (O(n ldr log(n))) duplication deletion algorithm for the case when closeness of two points (x1,y1) and (x2,y2) is described as closeness of both coordinates. In this paper, we extend this algorithm to the case when closeness is described by an arbitrary metric. Both algorithms have been successfully applied to gravity databases.","PeriodicalId":292853,"journal":{"name":"NAFIPS 2007 - 2007 Annual Meeting of the North American Fuzzy Information Processing Society","volume":"7 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2007-06-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Detecting Duplicates in Geoinformatics: from Intervals and Fuzzy Numbers to General Multi-D Uncertainty\",\"authors\":\"S. Starks, L. Longpré, R. Araiza, V. Kreinovich, Hung T. Nguyen\",\"doi\":\"10.1109/NAFIPS.2007.383900\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Geospatial databases generally consist of measurements related to points (or pixels in the case of raster data), lines, and polygons. In recent years, the size and complexity of these databases have increased significantly and they often contain duplicate records, i.e., two or more close records representing the same measurement result. In this paper, we address the problem of detecting duplicates in a database consisting of point measurements. As a test case, we use a database of measurements of anomalies in the Earth's gravity field that we have compiled. In our previous papers (2003,2004), we have proposed a new fast (O(n ldr log(n))) duplication deletion algorithm for the case when closeness of two points (x1,y1) and (x2,y2) is described as closeness of both coordinates. In this paper, we extend this algorithm to the case when closeness is described by an arbitrary metric. Both algorithms have been successfully applied to gravity databases.\",\"PeriodicalId\":292853,\"journal\":{\"name\":\"NAFIPS 2007 - 2007 Annual Meeting of the North American Fuzzy Information Processing Society\",\"volume\":\"7 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2007-06-24\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"NAFIPS 2007 - 2007 Annual Meeting of the North American Fuzzy Information Processing Society\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/NAFIPS.2007.383900\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"NAFIPS 2007 - 2007 Annual Meeting of the North American Fuzzy Information Processing Society","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/NAFIPS.2007.383900","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

地理空间数据库通常由与点(或光栅数据中的像素)、线和多边形相关的测量值组成。近年来,这些数据库的规模和复杂性显著增加,它们经常包含重复记录,即代表相同测量结果的两个或多个接近记录。在本文中,我们解决了在由点测量组成的数据库中检测重复的问题。作为一个测试案例,我们使用了一个我们编译的地球重力场异常测量数据数据库。在我们之前的论文(2003,2004)中,我们提出了一种新的快速(O(n ldr log(n)))重复删除算法,用于将两点(x1,y1)和(x2,y2)的接近度描述为两个坐标的接近度。在本文中,我们将该算法推广到用任意度量来描述接近度的情况。这两种算法都成功地应用于重力数据库。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Detecting Duplicates in Geoinformatics: from Intervals and Fuzzy Numbers to General Multi-D Uncertainty
Geospatial databases generally consist of measurements related to points (or pixels in the case of raster data), lines, and polygons. In recent years, the size and complexity of these databases have increased significantly and they often contain duplicate records, i.e., two or more close records representing the same measurement result. In this paper, we address the problem of detecting duplicates in a database consisting of point measurements. As a test case, we use a database of measurements of anomalies in the Earth's gravity field that we have compiled. In our previous papers (2003,2004), we have proposed a new fast (O(n ldr log(n))) duplication deletion algorithm for the case when closeness of two points (x1,y1) and (x2,y2) is described as closeness of both coordinates. In this paper, we extend this algorithm to the case when closeness is described by an arbitrary metric. Both algorithms have been successfully applied to gravity databases.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信