Upgrade to Pro — share decks privately, control downloads, hide ads and more …

Big Data and Bibliometrics: Crowdsourcing the W...

Big Data and Bibliometrics: Crowdsourcing the World’s Largest Open Database of Research

This presentation was delivered at O'Reilly's Strata conference in Santa Clara on how Mendeley is using technologies such as Hadoop and Mahout for scaling and recommendation.

William Gunn

March 01, 2012
Tweet

More Decks by William Gunn

Other Decks in Science

Transcript

  1. Big Data and Bibliometrics William Gunn Head of Academic Outreach

    [email protected] @mrgunn Crowdsourcing the World’s Largest Open Database of Research
  2. “The state of knowledge of the human race is sitting

    in the scientists’ computers, and is currently not shared […] We need to get it unlocked so we can tackle those huge problems.” A Big Problem
  3. http://www.plosmedicine.org/article/info:doi/10.1371/journal.pmed.0030291 Problems with Impact Factor “Thomson Scientific, the sole arbiter

    of the impact factor game...has no obligation to be accountable to... the authors and readers of scientific research. During discussions with Thomson Scientific over which article types in PLoS Medicine the company deems as “citable,” it became clear that the process of determining a journal's impact factor is unscientific and arbitrary.”
  4. Highly Tweeted articles are 11x more likely to be highly

    cited. (Eysenbach 2011) http://www.jmir.org/2011/4/e123/ The higher the impact factor, the more likely the research is to be retracted, partly due to intense competition. http://bjoern.brembs.net/news766.html.11 What matters is who is reading your work!
  5. ...and aggregates research data in the cloud Mendeley extracts research

    data… Mendeley makes science more collaborative and transparent: Install Mendeley Desktop Collecting rich signals from domain experts.
  6. 160 million documents uploaded 1.7 million users Cambridge Stanford University

    MIT Imperial College London University of Oxford Harvard University University of Michigan University College London University of California at Berkeley Columbia University The world's largest open database of research
  7. Big data problems we've had to solve Metadata extraction –

    We trained a 2-stage SVM to achieve precision at .91 and recall at .94, beating all other approaches. Deduplication – To build the web catalog, we've got to cluster and de-duplicate 17TB+ of documents daily. Author name disambiguation – aka the “big Wang problem”.We tried a variety of approaches settling on a method of hierarchical agglomerative clustering.
  8. Solving our big data problems We needed something that provided

    scalable processing as well as data storage, which made HDFS + MapReduce on AWS a pretty obvious choice. Stats Search good user experience enterprise-class search with easy setup vibrant open source community Scale SSDs for catalog search Caching index in RAM
  9. Solving our big data problems Deduplication PDFs are easier than

    pictures or audio because the descriptors are already text strings OCR doesn't work well Hashing works for trivial modifications, but not for discriminating pre-prints vs. post-prints. You don't necessarily know when you don't have a complete record. File hash check(SHA-1) Identifier check(e.g.PubMed id) Document fingerprint(fulltext) Metadata similarity check Update individual article page
  10. Mendeley was 3rd largest UK OpenURL referrer in April 2011,

    beating Medline and Scopus, with 34k click- throughs.
  11. http://dev.mendeley.com Tim O’Reilly O’Reilly Media James Powell CTO Thomson Reuters

    Juan Enriquez MD Excel Venture Management John Wilbanks VP Science, Creative Commons Werner Vogels CTO Amazon.com Mendeley/PLoS API Binary Battle $16,001 for the best app
  12. Select relation: supports refutes complements uses same method ... Result:

    A human-curated, constantly evolving semantic article database