reserved. GEnAI brings also new components and technologies … Proprietary API (OpenAI, Anthropic) Cloud Provider (AWS, GCP, Azure, COreweave) Open API (Hugging Face, Replicate) Opinionated Cloud (Databricks, Anyscale, Mosaic, Modal, Runpod,…) Contextual data Prompt Few-shot examples Query Output LLM APIs and Hosting LEGEND Gray boxes show key components of the stack, with leading tools/system listed Arrow show the flow of data through the stack Data Pipelines (Databricks, Airflow, Unstructured, …) APIs, Plugins (Serp, Wolfram, Zapier, …) LLM Cache (Redis, SQlite, GPTCache) Logging/LLMops (Weights & Biases, ML flow, PromptLayer, Helicone) Validation (Guardrails, Rebuff, Guidance, LMQL) Playground (OpenAI, nat.dev, Humanloop) App Hosting (Vercel, Steamship, Streamlit, Modal) Embedding Model (OpenAI, Cohere, Hugging Face) Orchestration (Phython/DIY, LangChain, LIamaIndex, ChatGPT) Vector Database (Pinecone, Weaviate, Chroma, pgvector) Output returned to users Queries submitted by users Prompts and few-shot examples that are sent to the LLM Contextual data provided by app developers to condition LLM output Source: Emerging Architectures for LLM Applications by Matt Bornstein and Rajko Radovanovic