a hierarchical way to extract image features ‣ Similar to how neurons in the visual cortex respond to particular stimuli CONVOLUTIONAL NEURAL NETWORKS (CNNs) 8
handle sequential data like text and speech – attention functionality allows the transformer to focus on specific parts of the input sequence ‣ Basis for Large Language Models (LLMs) ‣ Constructed to solve a specific problem, not to model a particular brain functionality TRANSFORMERS 10
could theoretically implement the core computations performed by transformers networks. ‣ Provides a novel perspective of the relationship between LLMs and the brain BRAIN- TRANSFORMERS? 11
the way it actually occurs – a model of how things actually are PHILOSOPHICAL ANALYSIS HOW-POSSIBLY MODEL ‣ Propositional model of how a phenomena might possibly occur – how things could possibly be 14
relationship is about similarities, about representing more or less, we cannot say: ‣ Model A represents a possibility that is actual ‣ Model B represents a possibility that is not actual THEY REPRESENT IN DEGREES AND RESPECT, NOT EITHER OR
DIVIDING MODELS INTO POSSIBLY–ACTUAL: ‣ Adjust their similarity requirements ‣ A model that succeeds in representing a target due to decreased similarity requirements should be viewed as a how-roughly model If we hold a model to less strict similarity requirements, it may succeed in representing a target, if only roughly
‣ We now this to be incorrect: ‣ An inaccurate, how-possibly model according to old view ‣ Shift to thinking about similarities: ‣ With decreased similarity requirements, the model would be correct ‣ A how-roughly model that captures important features, even if it is not 100% correct EXAMPLE: COPERNICAN MODEL
can be tested by running them in simulations where their performance is examined ‣ When we interact with ChatGPT, we are running the LLM in a simulation ‣ When we are using image recognition with CNNs, we run simulations that employ these models
reinforcement learning scenario where the agent behavior is studied in various ways ‣ Simulated ecosystem where behavior is learned through reinforcement learning RIGHT NOW
agents with LLMs ‣ Theory behind this is that the agents "cognitive functions" could be viewed as how-roughly models that are tested in an environment GOING FORWARD
the Copernican Model could learn us things about the solar system without being 100% correct, maybe transformers and LLMs can teach us about the brain without being 100% accurate models of the brain, or even intended as brain-models to begin with. They could be seen as how-roughly models. Simulated environments with embedded agents that employs transformer-based LLMs could help us to test this, and the same methodology could be used to test other cognitive theories.