The Fifth Elephant 2024 Annual Conference (12th &13th July)
Maximising the Potential of Data — Discussions around data science, machine learning & AI
Jul 2024
8 Mon
9 Tue
10 Wed
11 Thu
12 Fri
13 Sat 09:00 AM – 06:05 PM IST
14 Sun
r0
A lot of engineers are interested in using LLMs nowadays. However, its efficient execution remains a challenge. Efficient execution is key to mainstream adoption. To run them efficiently, we need accelerated systems such as GPU. This talk will explore the fundamentals of GPU architecture and its programming model, moving beyond model.to('cuda')
to understand the inner workings of GPUs. Attendees will gain insights into GPU internals and engage in a hands-on workshop using Triton, a programming language for writing efficient GPU code. We’ll cover core Triton concepts and implement essential operations for modern LLMs, empowering participants to optimize their models for better performance.
References
The ideal audience for my talk/workshop includes Data Scientists, ML engineers, and AI engineers aiming to enhance their model’s speed and maximize GPU usage. They are interested in learning Open AI’s Triton. Additionally, this session is suited for early-career researchers experimenting with their custom architectures and who want to improve the speed of their models.
Getting started with GPU programming can be challenging, especially with scatter resources. It makes it tough to understand how to start learning GPU programming. It is even tougher for Triton for which there are very limited resources. This talk aims to help the audience understand the basics of the GPU programming model and grasp the core concepts of Triton, making it easier to begin and improve their model speed. It will also cover the next steps in becoming better at GPU programming.
Here is how I am planning to split my talk:
The workshop will cover the basics of GPU hardware, programming models, and a hands-on walkthrough of a few kernels in Trition.
Participants will gain a deeper understanding of how GPUs work and learn to write performant code for GPUs. This knowledge will help them improve the throughput or reduce the latency of their existing models. The participants can then go ahead and implement their models from scratch, often rivaling HuggingFace’s implementation in latency and throughput.
We should have a workshop-based session lasting from 60-90 minutes. I would want to project a paper and pen to draw the actual computations in front of the audience and then present my IDE to code them live.
The coding session will be interactive and I would like it if people could code along with me.
Hosted by
Supported by
Gold Sponsor
Sponsor
Community Partner
Beverage Partner
{{ gettext('Login to leave a comment') }}
{{ gettext('Post a comment…') }}{{ errorMsg }}
{{ gettext('No comments posted yet') }}