We can't find the internet
Attempting to reconnect
Something went wrong!
Hang in there while we get back on track
BJ Hargrave- Open Source Community Instruction-tuning of Large Language Models | PyData Vermont 2024
Learn how InstructLab enables open-source community instruction tuning of LLMs using teacher-student models, synthetic data generation, and a structured taxonomy approach.
- 
    InstructLab is a new open-source project by IBM and Red Hat that enables community-driven instruction tuning of Large Language Models (LLMs) 
- 
    The project uses a “teacher-student” model approach: - Large teacher models generate synthetic training data
- Smaller student models learn from this data
- This reduces costs while maintaining capabilities
 
- 
    Contributions are organized in a taxonomy with two main types: - Knowledge recipes (facts and information)
- Compositional skill recipes (how to perform tasks)
 
- 
    The synthetic data generation process includes: - Question generation by teacher model
- Answer generation based on provided knowledge
- Quality evaluation of outputs
- Fine-tuning the student model with validated data
 
- 
    Key advantages of the approach: - Allows community contributions similar to open source software
- Addresses the problem of limited new training data
- Enables training on private/enterprise data
- More cost-effective than using large models like ChatGPT
 
- 
    Technical implementation details: - Runs on Apple Silicon Macs
- Uses 4-bit quantized models
- Supports GGUF format
- Available via pip install
- Compatible with models like Mistral and Llama
 
- 
    Project manages training quality through: - Replay buffers to prevent forgetting
- Even distribution of training across taxonomy
- Evaluation of synthetic data quality
- Balance between different skills and knowledge areas