We can't find the internet
Attempting to reconnect
Something went wrong!
Hang in there while we get back on track
LLMs gone wild - Tess Ferrandez-Norlander - NDC Oslo 2024
Learn about key components, challenges, and best practices for RAG systems with LLMs, including data preparation, accuracy metrics, and practical implementation tips.
-
RAG (Retrieval-Augmented Generation) systems are currently ~75% of LLM applications, helping ground LLM outputs in factual data
-
Key components of successful RAG implementations:
- Proper data preparation and chunking
- Selection of appropriate embedding models
- Effective metadata filtering
- Guardrails against hallucinations and sensitive data
- Evaluation frameworks for accuracy
-
LLMs have varying accuracy rates:
- Generation accuracy: 60-85%
- Retrieval accuracy: 60-85%
- Combined RAG system accuracy: ~72%
-
Critical optimization areas:
- Data ingestion and preparation
- Chunk size and strategy
- Context window management
- Prompt engineering
- Re-ranking of results
-
Common challenges:
- Handling sensitive/PII data
- Maintaining data freshness
- Dealing with multi-modal content (images, tables)
- Managing context windows
- Preventing hallucinations
-
Key evaluation metrics:
- Context precision
- Context recall
- Answer relevancy
- Factual accuracy
-
Best practices:
- Implement proper data filtering
- Use metadata enrichment
- Test with real users
- Add guardrails for sensitive use cases
- Monitor and evaluate system performance
-
Tools and frameworks:
- LangChain
- LlamaIndex
- Semantic Kernel
- Various vector databases
- Evaluation frameworks like Ragas