Nov 2024
18 Mon
19 Tue
20 Wed
21 Thu
22 Fri 09:00 AM – 05:10 PM IST
23 Sat
24 Sun
Nov 2024
18 Mon
19 Tue
20 Wed
21 Thu
22 Fri 09:00 AM – 05:10 PM IST
23 Sat
24 Sun
This video is for members only
LLMs and generative AI have made their way into our day-to-day operations. While the wrappers over GPT are a good starting point, I was intrigued by what it takes for an SRE to understand the domain, identify its operational aspects, and build runbooks around running self-hosted LLM models.
Currently, too many models are built, but very few are in production. While many companies are trying to streamline the toolchain, it is still nascent. The body of work I will discuss is an experiment to build an understanding of the LLMOps ecosystem.
We built an internal server setup and have explored deploying the models on GPUs instead of relying on OpenAI.
Goals:
What was our learning curve:
What will you gain from this talk:
This talk benefits software engineers at all levels but is especially relevant for SRE and DevOps practitioners.
Nov 2024
18 Mon
19 Tue
20 Wed
21 Thu
22 Fri 09:00 AM – 05:10 PM IST
23 Sat
24 Sun
Hosted by
Supported by
Platinum Sponsor
Platinum Sponsor
Community sponsor
Venue host - Rootconf workshops
Community Partner
Community Partner
{{ gettext('Login to leave a comment') }}
{{ gettext('Post a comment…') }}{{ errorMsg }}
{{ gettext('No comments posted yet') }}