Upgrade to Pro — share decks privately, control downloads, hide ads and more …

Real time data driven applications (and SQL vs NoSQL databases)

lanzani
November 22, 2014

Real time data driven applications (and SQL vs NoSQL databases)

These are the slides of the talk I gave at NoSQL Barcelona (November 2014)

lanzani

November 22, 2014
Tweet

More Decks by lanzani

Other Decks in Technology

Transcript

  1. GoDataDriven PROUDLY PART OF THE XEBIA GROUP Real time data

    driven applications Giovanni Lanzani Data Whisperer and SQL vs NoSQL databases
  2. Real-time, data driven app? •No store and retrieve; •Store, {transform,

    enrich, analyse} and retrieve; •Real-time: retrieve is not a batch process; •App: something your mother could use: SELECT attendees FROM NoSQLMatters WHERE password = '1234';
  3. Is it Big Data? Everybody talks about it Nobody knows

    how to do it Everyone thinks everyone else is doing it, so everyone claims they’re doing it… Dan Ariely
  4. Is it Big Data? •Raw logs are in the order

    of 40TB; •We use Hadoop for storing, enriching and pre- processing.
  5. •Harder than it looks; •Large data; •Retrieval is by giving

    date, center location + radius. 4. Real-Time Retrieval
  6. date hour id_activity postcode hits delta sbi 2013-01-01 12 1234

    1234AB 35 22 1 2013-01-08 12 1234 1234AB 45 35 1 2013-01-01 11 2345 5555ZB 2 1 2 2013-01-08 11 2345 5555ZB 55 2 2 Data Example
  7. date hour id_activity postcod e hits delta sbi 2013-01-01 12

    1234 1234AB 35 22 1 2013-01-08 12 1234 1234AB 45 35 1 2013-01-01 11 2345 5555ZB 2 1 2 2013-01-08 11 2345 5555ZB 55 2 2 Data Example
  8. helper.py example def get_statistics(data, sbi): sbi_df = data[data.sbi == sbi]

    # select * from data where sbi = sbi hits = sbi_df.hits.sum() # select sum(hits) from … delta_hits = sbi_df.delta.sum() # select sum(delta) from … if delta_hits: percentage = (hits - delta_hits) / delta_hits else: percentage = 0 return {"sbi": sbi, "total": hits, "percentage": percentage}
  9. helper.py example def get_timeline(data, sbi): df_sbi = data.groupby([“date”, “hour", “sbi"]).aggregate(sum)

    # select sum(hits), sum(delta) from data group by date, hour, sbi return df_sbi
  10. Who has my data? •First iteration was a (pre)-POC, less

    data (3GB vs 500GB); •Time constraints; •Oeps: everything is a pandas df!
  11. Advantage of “everything is a df” Pro: •Fast!! •Use what

    you know •NO DBA’s! •We all love CSV’s! Contra: •Doesn’t scale; •Huge startup time; •NO DBA’s! •We all hate CSV’s!
  12. •Set the dataframe index wisely; •Align the data to the

    index: •Beware of modifications of the original dataframe! source_data.sort_index(inplace=True) If you want to go down this path
  13. The reason pandas is faster is because I came up

    with a better algorithm If you want to go down this path
  14. Issues?! •With a radius of 10km, in Amsterdam, you get

    10k postcodes. You need to do this in your SQL: •Index on date and postcode, but single queries running more than 20 minutes. SELECT * FROM datapoints WHERE date IN date_array AND postcode IN postcode_array;
  15. PostGIS is a spatial database extender for PostgreSQL. Supports geographic

    objects allowing location queries: SELECT * FROM datapoints WHERE ST_DWithin(lon, lat, 1500) AND dates IN ('2013-02-30', '2013-02-31'); -- every point within 1.5km -- from (lat, lon) on imaginary dates Postgres + Postgis (2.x)
  16. How we solved it 1. Align data on disk by

    date; 2. Use the temporary table trick: 3. Lose precision: 1234AB→1234 CREATE TEMPORARY TABLE tmp (postcodes STRING NOT NULL PRIMARY KEY); INSERT INTO tmp (postcodes) VALUES postcode_array; SELECT * FROM tmp JOIN datapoints d ON d.postcode = tmp.postcodes WHERE d.dt IN dates_array;
  17. Take home messages 1. Geospatial problems are “hard” and cam

    kill your queries; 2. Not everybody has infinite resources: be smart and KISS! 3. SQL or NoSQL? (Size, schema)