We can't find the internet
Attempting to reconnect
Something went wrong!
Hang in there while we get back on track
Andrey Cheptsov - Leveraging open-source LLMs for production | PyData Global 2023
Learn how to leverage open-source LLMs in production with guidance on model selection, optimization techniques, fine-tuning approaches, and deployment considerations.
-
Open-source LLMs offer full control over model behavior, data processing, and privacy compared to proprietary models
-
Key advantages of open-source LLMs:
- Cost reduction potential through optimization
- Customization flexibility
- No vendor lock-in risk
- Full privacy control
- Community-driven improvements
-
Major open-source LLM models:
- LLaMA 2 (Meta/Microsoft) - available in 7B, 13B, 70B parameters
- Code LLaMA - specialized for code generation
- Mistral - efficient 7B model with strong performance
-
Technical optimization techniques:
- LoRA (Low Rank Adaptation) reduces memory requirements for training
- Quantization converts weights to lower precision formats (int8, int4)
- Combining LoRA and quantization can reduce memory needs by 10x
-
Fine-tuning approaches:
- Supervised fine-tuning for basic task adaptation
- RLHF (Reinforcement Learning from Human Feedback) for instruction following
- DPO (Direct Preference Optimization) as simpler alternative to RLHF
-
Production deployment considerations:
- Memory requirements vary significantly by model size
- Batch inference can improve latency
- Multiple models can be served using shared resources
- Commercial usage rights vary by model license
-
Benchmarks show open-source models approaching GPT-3.5/GPT-4 quality on specific tasks through fine-tuning
-
Common use cases:
- Code generation
- SQL query generation
- Structured data extraction
- API automation
- Synthetic dataset generation