LLM in Production Conference Takeaways
LLMs
Production
I didn’t get to attend the LLM in Production Conference but found these takeaways Demetrios Brinkmann shared in an email to be quite insightful:
- Data is still king - LLMs are great but if you don’t have quality clean data you won’t go far.
- Smaller models can be just as good as larger general models at specific tasks. And cheaper!
- Fine-tuning is becoming cheaper.
- Evaluation of LLMs is very hard - feels very subjective still.
- Managed APIs are expensive.
- “Traditional” ML isn’t going anywhere.
- Memory matters - for both serving and training.
- Information retrieval w/ vector databases is becoming standard pattern.
- Start w/ prompt engineering and push that to its limits before fine-tuning w/ smaller models.
- Use agents/chains only when necessary. They are unruly.
- Latency is critical for a good user experience.
- Privacy is critical.
As a practicing data scientist, #6 is reassuring!
Here are some of the videos: