17 April 2024 10:00 - 10:30
Navigating LLM deployment: tips, tricks, and techniques
Self-hosted Language Models are going to power the next generation of applications in critical industries like financial services, healthcare, and defence.
Self-hosting LLMs, as opposed to using API-based models, comes with its own host of challenges - as well as needing to solve business problems, engineers need to wrestle with the intricacies of model inference, deployment and infrastructure.
In this talk, Meryem will discuss the best practices in model optimisation, serving and monitoring - with practical tips and real case-studies.