Introduction and methodology

Regina Corso Consulting
{"title":"Introduction and methodology","authors":"Regina Corso Consulting","doi":"10.6027/9789289332453-3-en","DOIUrl":null,"url":null,"abstract":"©2009 EDUCAUSE. Reproduction by permission only. In the second decade of the Internet era, the world is awash in a stunning quantity and variety of information. Exactly how much digital information we’re producing and living alongside is a tough question to answer, but there’s no doubt that the answer involves the kinds of numbers we used to associate mainly with extragalactic distances. A study by the University of California, Berkeley, estimated that in 2002 a total of 5 exabytes (5 billion gigabytes) of new data was recorded on media (including paper)—37,000 times the size of the Library of Congress—and another 18 exabytes flowed through electronic channels.1 The IT research firm IDC, estimating the size of what the company calls the digital universe, produced an even more eye-popping figure: 281 exabytes of digital information created, captured, and replicated in 2007 alone, a quantity IDC expects to grow by 60% annually through 2011.2 These numbers reflect a particularly fast-growing component of image and video capture, but there seems to be hardly any endeavor that isn’t confronted with an exploding body of digital material. Scientific data captured in experiments or generated in computational research has reached heroic proportions, with no end in sight; at the upper end of data density, the Large Hadron Collider at CERN is expected to produce 15 million gigabytes of raw experimental data per year, a DVD every five seconds.3 The more bookish and text-bound side of the academy is undergoing a digital revolution of its own, as Google and other entities produce a searchable online corpus of books that could only have been found at the finest research libraries a decade ago—if there. Perhaps the most remarkable efflorescence of digital content has come from a combination of widening Internet access and web interactivity, a phenomenon almost impossible to quantify but suggested in outline by one estimate that almost 1.7 billion people worldwide now have access to the Internet, and another estimate that the number of available websites exceeds 225 million.4 And that doesn’t count more than 400,000 terabytes of new information generated annually by e-mail!5 Nor are more traditional sorts of structured business data lying still. In a 2007 study by the IT advisory firm Ventana Research, more than half of organizations surveyed reported that their structured business data was growing by 20% or more annually, and their unstructured content was growing faster still.6 For that matter, such neat categories as structured and unstructured data are getting intertwined, as back-end databases dynamically feed websites and XML-related technologies contribute to 2 Introduction and Methodology","PeriodicalId":217570,"journal":{"name":"Taxing Energy Use 2019","volume":"26 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"27","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Taxing Energy Use 2019","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.6027/9789289332453-3-en","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 27

Abstract

©2009 EDUCAUSE. Reproduction by permission only. In the second decade of the Internet era, the world is awash in a stunning quantity and variety of information. Exactly how much digital information we’re producing and living alongside is a tough question to answer, but there’s no doubt that the answer involves the kinds of numbers we used to associate mainly with extragalactic distances. A study by the University of California, Berkeley, estimated that in 2002 a total of 5 exabytes (5 billion gigabytes) of new data was recorded on media (including paper)—37,000 times the size of the Library of Congress—and another 18 exabytes flowed through electronic channels.1 The IT research firm IDC, estimating the size of what the company calls the digital universe, produced an even more eye-popping figure: 281 exabytes of digital information created, captured, and replicated in 2007 alone, a quantity IDC expects to grow by 60% annually through 2011.2 These numbers reflect a particularly fast-growing component of image and video capture, but there seems to be hardly any endeavor that isn’t confronted with an exploding body of digital material. Scientific data captured in experiments or generated in computational research has reached heroic proportions, with no end in sight; at the upper end of data density, the Large Hadron Collider at CERN is expected to produce 15 million gigabytes of raw experimental data per year, a DVD every five seconds.3 The more bookish and text-bound side of the academy is undergoing a digital revolution of its own, as Google and other entities produce a searchable online corpus of books that could only have been found at the finest research libraries a decade ago—if there. Perhaps the most remarkable efflorescence of digital content has come from a combination of widening Internet access and web interactivity, a phenomenon almost impossible to quantify but suggested in outline by one estimate that almost 1.7 billion people worldwide now have access to the Internet, and another estimate that the number of available websites exceeds 225 million.4 And that doesn’t count more than 400,000 terabytes of new information generated annually by e-mail!5 Nor are more traditional sorts of structured business data lying still. In a 2007 study by the IT advisory firm Ventana Research, more than half of organizations surveyed reported that their structured business data was growing by 20% or more annually, and their unstructured content was growing faster still.6 For that matter, such neat categories as structured and unstructured data are getting intertwined, as back-end databases dynamically feed websites and XML-related technologies contribute to 2 Introduction and Methodology
引言和方法论
EDUCAUSE©2009。仅经许可复制。在互联网时代的第二个十年,世界充斥着数量惊人、种类繁多的信息。确切地说,我们制造了多少数字信息,并与之共存,这是一个很难回答的问题,但毫无疑问,答案涉及到我们过去主要与河外距离联系在一起的各种数字。加州大学伯克利分校的一项研究估计,2002年总共有5eb(50亿千兆字节)的新数据被记录在媒体(包括纸张)上——是国会图书馆容量的37000倍——另外还有18eb通过电子渠道流动IT研究公司IDC估计了“数字宇宙”的规模,得出了一个更令人瞠目的数字:仅在2007年,就有281 eb的数字信息被创造、捕获和复制,IDC预计到2011年,这一数字将以每年60%的速度增长。这些数字反映了图像和视频捕获的一个特别快速增长的组成部分,但似乎几乎没有任何努力不面临爆炸式增长的数字材料。在实验中获得的科学数据或在计算研究中产生的科学数据已经达到了惊人的程度,而且看不到尽头;在数据密度的高端,欧洲核子研究中心的大型强子对撞机预计每年产生1500万千兆字节的原始实验数据,相当于每5秒生成一张DVD学术性和文本化程度更高的学术界正在经历一场自己的数字革命,b谷歌和其他实体制作了一个可搜索的在线图书库,而这些图书在十年前只有在最好的研究图书馆才能找到——如果有的话。也许数字内容最显著的繁荣是互联网接入和网络交互性的结合,这一现象几乎无法量化,但据估计,目前全球有近17亿人可以访问互联网,另一项估计显示,可用网站的数量超过2.25亿这还不包括每年通过电子邮件产生的超过40000tb的新信息!更传统的结构化商业数据也不是静止不动。在2007年IT咨询公司Ventana Research的一项研究中,超过一半的受访组织报告称,他们的结构化业务数据每年增长20%或更多,而他们的非结构化内容增长得更快就这一点而言,结构化和非结构化数据这类整洁的分类正变得越来越复杂,因为后端数据库动态地为网站提供信息,xml相关的技术也为2引言和方法论做出了贡献
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信