{"title":"Comparing bibliographic descriptions in seven free-access databases","authors":"Lorena Joaquina Delgado Quirós, J. Ortega","doi":"10.55835/6436c590b3340c364be5b2c7","DOIUrl":null,"url":null,"abstract":"This communication aims to analyse the information that a large set of free-access databases (i.e., Crossref, Dimensions, Microsoft Academic, OpenAlex, Scilit, Semantic Scholar, The Lens) provides about indexed publications in their databases. Using a random sample of 116k publications from Crossref, each database was queried to retrieve the same document list with the purpose of comparing the metadata of their publications. The results show that the completeness degree is different between databases and that the search engines show more problems to extract abstracts and assign document typologies. Dimensions is the product that obtain the highest completeness percentages in abstracts, open access documents, bibliographic data and document types.","PeriodicalId":334841,"journal":{"name":"27th International Conference on Science, Technology and Innovation Indicators (STI 2023)","volume":"4 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-05-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"27th International Conference on Science, Technology and Innovation Indicators (STI 2023)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.55835/6436c590b3340c364be5b2c7","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
This communication aims to analyse the information that a large set of free-access databases (i.e., Crossref, Dimensions, Microsoft Academic, OpenAlex, Scilit, Semantic Scholar, The Lens) provides about indexed publications in their databases. Using a random sample of 116k publications from Crossref, each database was queried to retrieve the same document list with the purpose of comparing the metadata of their publications. The results show that the completeness degree is different between databases and that the search engines show more problems to extract abstracts and assign document typologies. Dimensions is the product that obtain the highest completeness percentages in abstracts, open access documents, bibliographic data and document types.
本交流旨在分析大型免费数据库(即Crossref, Dimensions, Microsoft Academic, OpenAlex, Scilit, Semantic Scholar, the Lens)提供的关于其数据库中索引出版物的信息。使用来自Crossref的116k出版物的随机样本,查询每个数据库以检索相同的文档列表,以便比较其出版物的元数据。结果表明,不同数据库的检索结果完备程度不同,检索引擎在摘要提取和文档类型分配等方面存在较多问题。维度是在摘要、开放获取文档、书目数据和文档类型中获得最高完整性百分比的产品。