Slide 1

Slide 1 text

Aron Walsh Department of Materials Centre for Processable Electronics Machine Learning for Materials 8. Accelerated Discovery

Slide 2

Slide 2 text

Module Contents 1. Introduction 2. Machine Learning Basics 3. Materials Data 4. Crystal Representations 5. Classical Learning 6. Deep Learning 7. Building a Model from Scratch 8. Accelerated Discovery 9. Generative Artificial Intelligence 10. Future Directions

Slide 3

Slide 3 text

Every step balances system exploration and exploitation Key Concept #8 Policy Environment Reward Use data to choose the next best experiment (or simulation) Action Feedback Observation Example: “Choose next anneal temperature to maximise conductivity”

Slide 4

Slide 4 text

“A problem in artificial intelligence is one which is so complex that it cannot be solved using any normal algorithm” Hugh M. Cartwright, Applications of AI in Chemistry (1993)

Slide 5

Slide 5 text

Class Outline Accelerated Discovery A. Automated Experiments B. Bayesian Optimisation C. Reinforcement Learning

Slide 6

Slide 6 text

Accelerate Scientific Discovery Research can be broken down into a set of tasks that can each benefit from acceleration H. S. Stein and J. M. Gregoire, Chem. Sci. 10, 9640 (2019) Traditional research workflow

Slide 7

Slide 7 text

Potential for speedup Accelerate Scientific Discovery H. S. Stein and J. M. Gregoire, Chem. Sci. 10, 9640 (2019) Research can be broken down into a set of tasks that can each benefit from acceleration

Slide 8

Slide 8 text

Accelerate Scientific Discovery Workflow classification of published studies H. S. Stein and J. M. Gregoire, Chem. Sci. 10, 9640 (2019)

Slide 9

Slide 9 text

Automation and Robotics Execution of physical tasks to achieve a target using autonomous or collaborative robots Image: https://www.thefifthindustrialrevolution.co.uk

Slide 10

Slide 10 text

Automation and Robotics Robots can be tailored for a wide range of materials synthesis and characterisation tasks B. P. MacLeod et al, Science Advances 6, eaaz8867 (2020)

Slide 11

Slide 11 text

Automation and Robotics Self-driving labs (SDL) are now operating N. J. Szymanski et al, Nature 624, 86 (2023); Phys. Rev. Energy 3, 011002 (2024) A-Lab, Berkeley

Slide 12

Slide 12 text

Flexible Automation Systems Modular hardware with computer-controlled synthesis and characterisation DIGIBAT: In collaboration with Magda Titirici, Ifan Stephens, and others (ICL)

Slide 13

Slide 13 text

Flexible Automation Systems Automation platforms designed to deliver complex research workflows (fixed platform or mobile) DIGIBAT: In collaboration with Magda Titirici, Ifan Stephens, and others (ICL) Usually a mix of proprietary code, with GUI and Python API for user control

Slide 14

Slide 14 text

Automation and Robotics Robots can be equipped with sensors and artificial intelligence to interact with their environment S. Eppel et al, ACS Central Science 6, 1743 (2020) Adapting computer vision models for laboratory settings GT = ground truth Pred = predicted

Slide 15

Slide 15 text

Automation and Robotics Robots can be equipped with sensors and artificial intelligence to interact with their environment https://www.youtube.com/watch?v=K7I2QJcIyBQ

Slide 16

Slide 16 text

Optimisation Algorithms to efficiently achieve a desired research objective. Considerations: Objective function (O): Materials properties or device performance criteria, e.g. battery lifetime Parameter selection: Variables that can be controlled, e.g. temperature, pressure, composition Data acquisition: How the data is collected, e.g. instruments, measurements, automation

Slide 17

Slide 17 text

Optimisation Algorithms Local optimisation – find the best solution in a limited region of the parameter space (x) Gradient based: iterate in the direction of the steepest gradient (∇O), e.g. gradient descent Hessian based: use information from the second derivatives (∇2O), e.g. quasi-Newton O x x1 xn Local minimum The same concepts are involved in ML model training

Slide 18

Slide 18 text

Optimisation Algorithms Global optimisation – find the best solution from across the entire parameter space Numerical: iterative techniques to explore parameter space, e.g. downhill simplex, simulated annealing Probabilistic: incorporate probability distributions, e.g. Markov chain Monte Carlo, Bayesian optimisation O x Global minimum xn x1 The same concepts are involved in ML model training

Slide 19

Slide 19 text

Class Outline Accelerated Discovery A. Automated Experiments B. Bayesian Optimisation C. Reinforcement Learning

Slide 20

Slide 20 text

Bayesian Optimisation (BO) Use prior (measured or simulated) data to decide which experiment to perform next J. Močkus, Optimisation Techniques 1, 400 (1974) Probabilistic (Surrogate) Model Approximation of the true objective function O(x) ~ f(x), e.g. Gaussian process, GP(x,x') Acquisition Function Selection of the next sample point, e.g. upper confidence bound (UCB), probability of improvement (PI), expected improvement (EI) known new (parameters to sample)

Slide 21

Slide 21 text

Bayesian Optimisation (BO) Use prior (measured or simulated) data to decide which experiment to perform next J. Močkus, Optimisation Techniques 1, 400 (1974) Probabilistic (Surrogate) Model (parameters to sample) Gaussian process: f(x) ~ GP(μ(x), k(x,xʹ)) mean function Gaussian kernel function k(x,xʹ) measures the similarity points x and xʹ • Kernel controls function smoothness and uncertainty • Similar x share information • Dissimilar x default to the mean with high uncertainty known new

Slide 22

Slide 22 text

Bayesian Optimisation (BO) Bayesian optimisation for chemistry: Y. Wu et al, Digital Disc. 3, 1086 (2024) Use prior (measured or simulated) data to decide which experiment to perform next

Slide 23

Slide 23 text

Upper confidence bound (UCB) trades off mean prediction and uncertainty Exploration–Exploitation Tradeoff xnext = max( μ(x) + βσ(x) ) Prediction based on prior knowledge Weighted Uncertainty What to do next N. Srinivas et al, IEEE Transactions on Information Theory, 58 (2012) x β < 1 focus on exploitation β ~ 1 balance risk and reward β > 1 focus on exploration A tunable (learnable) hyperparameter of UCB

Slide 24

Slide 24 text

Applications of BO Maximise electrical conductivity of a composite (P3HT-CNT) thin-film D. Bash et al, Adv. Funct. Mater. 31, 2102606 (2021)

Slide 25

Slide 25 text

Applications of BO D. Bash et al, Adv. Funct. Mater. 31, 2102606 (2021) Maximise electrical conductivity of a composite (P3HT-CNT) thin-film

Slide 26

Slide 26 text

Active Learning (AL) BO: find inputs that maximise the objective function AL: find inputs that minimise uncertainty Epistemic uncertainty* Posterior samples Target unknown regions that can improve the model Gaussian process is updated with new observations to yield revised function values and uncertainty *Epistemic = data-limited (reducible) uncertainty; Aleatoric = noise (irreducible)

Slide 27

Slide 27 text

Integrated Research Workflows Feedback loop between optimisation model and automated experiments NIMS-OS: R. Tamura, K. Tsuda, S. Matsuda, arXiv:2304.13927 (2023)

Slide 28

Slide 28 text

Integrated Research Workflows Feedback loop between optimisation model and automated experiments NIMS-OS: R. Tamura, K. Tsuda, S. Matsuda, arXiv:2304.13927 (2023)

Slide 29

Slide 29 text

Class Outline Accelerated Discovery A. Automated Experiments B. Bayesian Optimisation C. Reinforcement Learning

Slide 30

Slide 30 text

Reinforcement Learning (RL) Boston Dynamics Early applications in video games (maximise score), finance (profit), and robotics (task completion) Nintendo An agent interacts with an environment to learn decision-making strategies that achieve a specific goal

Slide 31

Slide 31 text

Reinforcement Learning (RL) RL schematic: S. J. D. Prince, https://udlbook.github.io

Slide 32

Slide 32 text

Reinforcement Learning (RL) Digital lab RL schematic: S. J. D. Prince, https://udlbook.github.io/udlbook Virtual scientist New experiment Property change Composition, structure, processing Design strategy (explore/exploit)

Slide 33

Slide 33 text

RL Policy This familiar equation is a softmax (Boltzmann) policy Data-driven decision making that adapts over time Probability of action at given state st Expected reward for action at Effective temperature for exploration/exploitation balance Sum over all possible actions Q(s,a) is the expected reward from taking action a in state s, following policy π

Slide 34

Slide 34 text

RL of Metal-Organic Frameworks This familiar equation is a softmax (Boltzmann) policy Hyunsoo Park et al, Digital Discovery 3, 728 (2024) Selectivity: 𝑆𝐶𝑂2/𝐻2𝑂 = 𝐾𝐻,𝐶𝑂2 𝐾𝐻,𝐻2𝑂 Heat of adsorption : 𝑄𝑠𝑡

Slide 35

Slide 35 text

RL of Metal-Organic Frameworks Hyunsoo Park et al, Digital Discovery 3, 728 (2024)

Slide 36

Slide 36 text

RL of Metal-Organic Frameworks Hyunsoo Park et al, Digital Discovery 3, 728 (2024)

Slide 37

Slide 37 text

Experimental Optimisation Strategies Advantages Disadvantages High- throughput (Enumeration) - Exhaustive search - Simple to implement and understand - Inefficient for high-dimensional spaces - Maximises experiments and dataset size Bayesian Optimisation - Efficiently exploit data - Works with noisy and expensive evaluations - Surrogate model & acquisition selection - Struggle with high- dimensional spaces Reinforcement Learning - Learns optimal policies through feedback - Can handle dynamic and complex environments - Sample hungry - Slow to converge

Slide 38

Slide 38 text

Obstacles to Closed Loop Discovery • Materials complexity: complex structures, compositions, processing sensitivity • Data quality and reliability: errors and inconsistencies waste resources • Cost of automation: major investment required in infrastructure and training • Adaptability: systems and workflows may be difficult to reconfigure for new problems

Slide 39

Slide 39 text

Class Outcomes 1. Select an appropriate optimisation strategy for a given problem 2. Assess the impact of AI optimisation tools on materials research and discovery Activity: Closed-loop optimisation