Archive for the ‘02: Analysis’ Category
A great article by the Webtrends CEO:
“We are living in “the age of big data, (…) No industry is untouched by big data, which is notably transforming the way social networks work today. However, the key factor that will determine success for companies in this age is not simply big data, but big science. (…)
The World Economic Forum’s report on data equated it with an asset such as gold. Others have declared that data is “the new oil.” But, as with gold or oil, data has no intrinsic value. (…) Gold requires mining and processing before it finds its way into our jewelry (…) Oil requires extraction and refinement before it becomes the gasoline that fuels our vehicles. Likewise, data requires collection, mining and, finally, analysis before we can realize its true value for businesses, governments, and individuals alike.
According to IDC, the amount of data that companies are wrestling with is growing at 50 percent per year — or more than doubling every two years. Many organizations are rich in data but poor in insight. That’s where big science comes in. (…)
The collection and mining of massive amounts of digital data currently defines the term big data. Those are processes that businesses largely handle. However, the analysis of that data — that magic ingredient of algorithms and advanced mathematics that bridges the gap between knowledge and insight — is big science. It is where the value is. It is the future. (…)
Put simply, the analysis that big science brings to the table makes big data relevant. I envision big science combining with big data to create big opportunities in three significant ways: real-time relevant content, data visualization, and predictive analytics. (…)
Read the full article at:
Yoder, Alex. Big data is worth nothing without Big science. C/Net, May 15, 2012. Available from: http://news.cnet.com/8301-1001_3-57434736-92/big-data-is-worth-nothing-without-big-science/ [Accessed on 15th of May 2012]
An impressive study research, sponsored by the Publishing Research Consortium: includes interviews of key-people from Pfizer, the CERN, Mendeley, the British Library, from TEMIS, Elsevier, Springer, Nature, Wiley, etc.
Journal Article Mining: a research study into Practices, Policies, Plans …..and Promises Eefke Smit and Maurits van der Graaf. PRC June 2011 153pp. This is a study commissioned by PRC which offers the first comprehensive look at what publishers and others are doing, and plan to do, in both data and text mining of the scholarly, mainly journal, literature. Lots of fascinating detail from a number of viewpoints – from 29 interviews and 190 detailed responses to a survey
This toolkit is designed and offered by the JISC to provide a guide to measuring the impacts of online scholarly resources.
It 1will help content creators, publishers and other information professionals understand the reach of their digital assets.
They can use the kit to help guide them through different aspects of measuring impact, both qualitative, such as focus groups, and quantitative, such as web metrics.
With his world map of scientific collaboration, Oliver Beauchesne, from the US-Canada based Science-Metrix, has built nice visualizaion of science collaboration.
It is based on Scopus data. For those interested in looking at how scientists are connected geographically, a number of companies already promise to help map the geographic reach of an individual or a discipline. These include Springer’s AuthorMapper, Transinsight’s GoPubMed and BioMedExperts.
Van Noorden, Richard. Picture post: world map of scientific collaboration. The Great Beyond (Nature), Posted on January 27,2011.
BioSumm is a flexible and modular framework which analyzes large collections of unclassified biomedical texts and produces ad hoc summaries oriented to biological information.
BioSumm is neither a traditional summarizer nor a extractor of dictionary terms. It is designed to be a summarizer oriented to the biological domain. Thus, its summaries have both the expressive power of the traditional summaries and the domain specificity of documents produced by a dictionary entry extractor.
The GUI interface is freely downloadable:
Update (31/01/2011): New address:
STM publisher Springer has announced the launch of a new free analytics tool, http://realtime.springer.com, which provides multiple visualisations of the usage that is generated worldwide by springer’s online products, including journals, books, images and protocols.
Realtime.springer.com aggregates the raw data on downloads of Springer journal articles and book chapters in real time from all over the world, and displays them in a variety of interactive visualisations such as:
- a map showing where the downloads are coming from
- a constantly updating keyword tag cloud
- and a visualisation of total downloads.
In addition, a search feature shows a chart of the downloads and the ‘Top Five Most Downloaded’ list for every journal or book.
The results provide book authors and journal editors with information on how intensively their content is used.
They gain insight into what topics are trending at the moment, and which areas of the world are currently looking at what type of topics in Springer books and journals.
Librarians get a clear overview of where Springer content is used in the many fields.
Realtime.springer.com currently receives input from the information platform SpringerLink with nearly five million documents from about 41,000 eBooks, 1,160 book series, 2,524 journals and 173 eReference works. Additionally, the tool receives feeds from the SpringerImages database with more than 2.7 million images and from SpringerProtocols, the database of reproducible laboratory protocols in the life and biomedical sciences