Large Language Models (LLMs) have enormous potential, but also challenge existing workflows in industry that require modularity, transparency and structured data. In this talk, I'll present pragmatic and practical approaches for how to use the latest generative models beyond just chat bots. I'll dive deeper into spaCy's LLM integration, which lets you plug in open-source and proprietary models and provides a robust framework for extracting structured information from text, distilling large models into smaller task-specific components, and closing the gap between prototype and production.
https://spacy.io/usage/large-language-models
The spacy-llm package integrates LLMs into spaCy pipelines, featuring a modular system for fast prototyping and prompting, and turning unstructured responses into robust outputs for various NLP tasks.
https://explosion.ai/blog/human-in-the-loop-distillation
This blog post presents practical solutions for using the latest state-of-the-art models in real-world applications and distilling their knowledge into smaller and faster components that you can run and maintain in-house.
https://prodi.gy/docs/large-language-models
Prodigy comes with preconfigured workflows for using LLMs to speed up and automate annotation and create datasets for distilling large generative models into more accurate, smaller, faster and fully private task-specific components.