Upgrade to Pro — share decks privately, control downloads, hide ads and more …

Using GPUs for Lightning Fast Analytics on MapD

Using GPUs for Lightning Fast Analytics on MapD

From MaptimeLA: https://www.meetup.com/MaptimeLA/events/250176047/

For our next MaptimeLA event, we'll learn about MapD - a powerful platform for visualizing large datasets.

GPU-powered in-memory databases and analytics platforms are the logical successors to CPU in-memory systems, largely due to recent increases in the onboard memory available on GPUs. With sufficient memory, GPUs possess numerous advantages over CPUs, including much greater compute and memory bandwidth, as well as a native graphics pipeline for visualization.

In this tutorial, Aaron Williams, VP of Community at MapD, will demo how MapD is able to leverage multiple GPUs per server to extract orders-of-magnitude performance increases over CPU-based systems, bringing interactive querying and visualization to multi-billion (with a ‘b’) row datasets.

OmniSci

May 10, 2018
Tweet

More Decks by OmniSci

Other Decks in Technology

Transcript

  1. © MapD 2018 Aaron Williams VP of Global Community @_arw_

    [email protected] /in/aaronwilliams/ /williamsaaron slides: https://speakerdeck.com/mapd
  2. © MapD 2018 © MapD 2018 3 “Every business will

    become a software business, build applications, use advanced analytics and provide SaaS services.” - Smart CEO Guy has
  3. © MapD 2018 The Evolution of Competitive Data 4 Collect

    It Make It Actionable Make it Predictive
  4. Core Density Makes a Huge Difference 5 GPU Processing CPU

    Processing 40,000 Cores 20 Cores *fictitious example Latency Throughput CPU 1 ns per task (1 task/ns) x (20 cores) = 20 tasks/ns GPU 10 ns per task (0.1 task per ns) x (40,000 cores) = 4,000 task per ns Latency: Time to do a task. | Throughput: Number of tasks per unit time.
  5. © MapD 2018 Advanced memory management Three-tier caching to GPU

    RAM for speed and to SSDs for persistent storage 8 SSD or NVRAM STORAGE (L3) 250GB to 20TB 1-2 GB/sec CPU RAM (L2) 32GB to 3TB 70-120 GB/sec GPU RAM (L1) 24GB to 256GB 1000-6000 GB/sec Hot Data Speedup = 1500x to 5000x Over Cold Data Warm Data Speedup = 35x to 120x Over Cold Data Cold Data COMPUTE LAYER STORAGE LAYER Data Lake/Data Warehouse/System Of Record
  6. © MapD 2018 Coming Soon 9 LIDAR Data • Working

    with Michael Flaxman (Geodesign Technologies) • An open dataset from the state of Florida with several billion points National Water Model • Working with Harvard Center for Geographic Analysis (Josh Lieberman and Ben Lewis) Open Source MapD Core Working Group Forming Soon • New features coming soon to the open source SQL engine • If you’re interested in participating, talk with the MapD team
  7. © MapD 2018 © MapD 2018 • mapd.com/demos Play with

    our demos • mapd.cloud Get a MapD instance in less than 60 seconds • mapd.com/platform/download-community/ Download the Community Edition • community.mapd.com Ask questions and share your experiences 10 Next Steps
  8. © MapD 2018 Aaron Williams VP of Global Community @_arw_

    [email protected] /in/aaronwilliams/ /williamsaaron slides: https://speakerdeck.com/mapd Thank you! Any questions?