Slide 14
Slide 14 text
13
Copyright © 2023 by Boston Consulting Group. All rights reserved.
GEnAI brings also new components and technologies …
Proprietary API
(OpenAI, Anthropic)
Cloud Provider
(AWS, GCP, Azure,
COreweave)
Open API
(Hugging Face, Replicate)
Opinionated Cloud
(Databricks, Anyscale,
Mosaic, Modal, Runpod,…)
Contextual data
Prompt
Few-shot examples
Query
Output
LLM APIs and Hosting
LEGEND
Gray boxes show key components of the stack, with leading tools/system listed
Arrow show the flow of data through the stack
Data Pipelines
(Databricks, Airflow,
Unstructured, …)
APIs, Plugins
(Serp, Wolfram, Zapier, …)
LLM Cache
(Redis, SQlite, GPTCache)
Logging/LLMops
(Weights & Biases, ML flow,
PromptLayer, Helicone)
Validation
(Guardrails, Rebuff,
Guidance, LMQL)
Playground
(OpenAI, nat.dev,
Humanloop)
App Hosting
(Vercel, Steamship,
Streamlit, Modal)
Embedding Model
(OpenAI, Cohere,
Hugging Face)
Orchestration
(Phython/DIY, LangChain,
LIamaIndex, ChatGPT)
Vector Database
(Pinecone, Weaviate,
Chroma, pgvector)
Output returned to users
Queries submitted by users
Prompts and few-shot examples that are sent to the LLM
Contextual data provided by app developers to condition LLM output
Source: Emerging Architectures for LLM Applications by Matt Bornstein and Rajko Radovanovic