Upgrade to Pro — share decks privately, control downloads, hide ads and more …

Securing and Protecting Your Data in the AI Era

Securing and Protecting Your Data in the AI Era

Avatar for Karen Lopez

Karen Lopez

May 23, 2025
Tweet

More Decks by Karen Lopez

Other Decks in Technology

Transcript

  1. Karen Lopez Microsoft MVP, Data Platform Microsoft Certified Trainer, vExpert

    Data management expert, space enthusiast, and #TeamData evangelist www.datamodel.com @datachick.bsky.social
  2. 5

  3. What is new? – Adversarial ML (AML) Generative AI generally

    connects to data, records, and databases, providing yet another vector to our data Machine Learning processes give attackers more avenues to manipulate and abuse data Vendors typically do not share what data has been used or will be used to train models 9
  4. Great Focus on Security, Privacy, and Compliance Data privacy is

    becoming increasingly important, with regulations placing greater emphasis on data protection. There is a growing need for data security due to increasing cyber attacks and data breaches. Compliance with regulations and standards is crucial for avoiding legal and financial penalties.
  5. More focus on AI and Data Artificial intelligence (AI) is

    growing in popularity as a tool for data analysis. Machine learning, a subset of AI, is used to make predictions and decisions based on data. Neural networks, another subset of AI, apply pattern recognition to data. There is a growing need for ethical considerations and responsible use of data. 12
  6. AI Security Concerns – Generative AI Data Bias Data Misunderstanding

    Social Engineering / Fakes Compliance Transparency / Validation
  7. Training • Data (Training and Testing) • Labels • Parameters

    • Code Model Deployment • Evasion • Privacy Predictive AI Exposure Points
  8. Targeted Poisoning Attackers manipulate subsets of training data to influence

    the model to make specific incorrect predictions.
  9. Backdoor Poisoning Adding a tiny change that humans can’t easily

    detect and then using that poison on future data. Examples include objects, reflections, and small triggers
  10. Prompt Stealing Taking carefully crafted prompts and using them without

    permission or asking the model to share previous prompts and instructions
  11. Prompt Injection Ask the model to give you the data

    or telling it to ignore previous instructions
  12. Jailbreaks and Role- playing Instructions Asking the model to respond

    as another persona to bypass guardrails or to otherwise break its rules
  13. NIST on AI Manipulation Evasion Attacks Poisoning Attacks Privacy Attacks

    Abuse https://www.nist.gov/news-events/news/2024/01/nist-identifies-types- cyberattacks-manipulate-behavior-ai-systems Predictive AI Generative AI
  14. Demands for Better Data Trust Data Catalog Data Governance Data

    Security Transparent Policies Data Contract Continuous Monitoring 31
  15. Data an AI Success Challenges Data Literacy Data Ethics Tools

    Not Keeping Up Changing Data Processes
  16. Data and AI Opportunities Using AI for Data Management Data-driven

    Projects Will Increase Demand Increased Focus on Data Quality Extend Data Management Capabilities
  17. No time today, but.. We should certainly be looking at

    how to use AI tools and techniques to protect data, just like any other tools We’d need to secure those AI systems from the same things we are talking about today All data protection best practices from non-AI systems still apply.
  18. Adversarial Training Training a model on both clean and adversarial

    data in order to increase the model's robustness.
  19. Randomized Smoothing Adds random noise to training data in order

    to increase the model's robustness by averaging out bad data
  20. The process of selecting and preparing data for use in

    machine learning models. It includes continuous monitoring and protection of that data Training Data Curation
  21. Label Monitoring Monitoring the accuracy of labels in a machine

    learning dataset, identifying mislabeled data and correcting the labels.
  22. Trigger Identification Triggers can be words, phrases, or other features

    that cause a model to produce incorrect predictions.
  23. Model Inspection Model inspection is important for understanding how a

    model is making predictions and identifying potential attacks and manipulations
  24. Supply Chain Security Ensuring the security of all components of

    a machine learning system: code, data, configurations, models
  25. Takeaways 1. Data protection in AI opens more attack surfaces

    2. Data may be at risk in ways most have not thought of 3. We can use AI to help us secure data 49 Training data Models Prompts Instructions Properties and Attributes