Skip to content
Exploring Machine Learning, AI, and Data Science

Ronen Dar on GPU Orchestration for Building ML Models

In this episode, our Andy Leonard and Frank La Vigne sit down with Ronen Dar, the co-founder and CTO of Run AI, to explore the world of artificial intelligence and GPU orchestration for machine learning models.

Ronen shares insights into the challenges of utilizing GPUs in AI research and how Run AI’s platform addresses these issues by optimizing GPU usage and providing tools for easier and faster model training and deployment. The conversation delves into the concept of fractional GPU usage, allowing multiple workloads to run on a single GPU, making expensive GPUs more accessible and cost-effective for organizations.

Links

Show Notes

04:40 GPU technology enabled for cloud AI workloads.

07:00 RunAI enables sharing expensive GPU resources for all.

11:59 As enterprise AI matures, organizations become more savvy.

15:35 Deep learning, GPUs for speed, CPUs backup.

16:54 LLMs running on GPU’s, exploding in market.

23:29 NVIDIA created CUDA to simplify GPU use.

26:21 NVIDIA’s success lies in accessible technology.

28:25 Solve GPU hugging with quotas and sharing.

31:15 Team lead manages GPU quotas for researchers.

35:51 Rapid changes in business and innovation.

40:34 Passionate problem-solver with diverse tech background.

43:38 Thanks for tuning in, subscribe and review.

Transcript
Speaker:

Greetings, listeners. Welcome back to the Data

Speaker:

Driven Podcast. I'm Bailey, your AI host with

Speaker:

the most data, that is, bringing you insights from the ether

Speaker:

with my signature wit. In today's episode, we're

Speaker:

diving deep into the heart of artificial intelligence's engine room,

Speaker:

GPU orchestration. It's the unsung hero

Speaker:

of AI research, optimizing the raw power needed to fuel

Speaker:

today's most advanced machine learning models. And

Speaker:

who better to guide us through this labyrinth of computational complexity than

Speaker:

Ronan Darr, the cofounder and CTO of Run AI, the

Speaker:

company that's making GPU resources work smarter, not

Speaker:

harder. Now onto the show.

Speaker:

Hello, and welcome to Data Driven, the podcast where we For the emergent fields

Speaker:

of artificial intelligence, data engineering, and overall data

Speaker:

science and analytics. With me as always is my favoritest

Speaker:

Data engineer in the world, Andy Leonard. How's it going, Andy? It's

Speaker:

going well, Frank. How are you? I'm doing great. I'm doing great. It's been,

Speaker:

we're We're recording this February 1, 2024. And as I said to my

Speaker:

kids yesterday, January has been a long year.

Speaker:

We're only, like, 1 month into the year, and it was it was a pretty

Speaker:

wild ride. But I can tell we're gonna have a blast today,

Speaker:

because we're gonna geek out on something that I kinda sort of understand,

Speaker:

but not entirely, and it's GPUs. And in the virtual green room, were chit

Speaker:

chatting with some folks, and, but let me do the formal introduction

Speaker:

here. Today with us, we have doctor Ronadhar, cofounder and CTO

Speaker:

of Run AI, A company at the forefront of GPU

Speaker:

orchestration, and he has a distinguished career in technology.

Speaker:

His experience includes significant roles at Apple. Yes, That

Speaker:

apple. Bell Labs. Yes. That Bell Labs.

Speaker:

And at Run AI, Ronan is instrumental in optimizing

Speaker:

GPU usage For AI model training and deployment,

Speaker:

leveraging his deep passion for both academia and startups.

Speaker:

And, Run AI is a key player in the, and he is a he

Speaker:

and Run AI are key player in the AI revolution. Ronan's

Speaker:

contribute Contributions are pivotable in shaping and powering the

Speaker:

future of artificial intelligence. Now I will add that in

Speaker:

my day job at Red Hat, Run AI has come up a couple of times.

Speaker:

So this is definitely, definitely

Speaker:

an honor to have you on on on the show, sir. Welcome.

Speaker:

Thank you, Frank. Thank you for inviting me. Hey, Andy. Good to

Speaker:

be here. I love it. Love Reddit. We're a big

Speaker:

fan of Reddit. We're working closely with many people in

Speaker:

Reddit, and love that. Right? Love OpenShift,

Speaker:

love Reddit, love Linux. Yeah. Cool. Cool.

Speaker:

Yeah. So so for those who don't know exactly, I kinda know

Speaker:

what, your Run AI does, but can you explain exactly

Speaker:

What it is run AI does and why GPU

Speaker:

orchestration is important. Yes.

Speaker:

Okay.

Speaker:

So run AI is, software,

Speaker:

AI infrastructure platform. So we

Speaker:

help machine learning teams to get much more

Speaker:

out of their GPUs, And we provide

Speaker:

those teams with abstraction layers and tools

Speaker:

so they can train models And deploy models

Speaker:

much easier, much faster. And

Speaker:

so We started in 2018, 6 years

Speaker:

ago. It's me and my cofounder, Omuri. Omuri is the CEO.

Speaker:

He's, he's amazing. I love him. We We know each other for many

Speaker:

years. We we met in the academia, like, more than 10 years ago,

Speaker:

and and we started running AI together, and We started

Speaker:

running AI because we saw that there are big challenges

Speaker:

around, GPU's, around orchestrating

Speaker:

GPU's and utilizing GPU's. We saw back then

Speaker:

in 2018, the GPUs are going to be very very important.

Speaker:

It's like the basic a a component in

Speaker:

that any AI company need to train models,

Speaker:

right, and deploy models. So we saw that GPUs are going to be critical, but

Speaker:

there are also a lot of challenges with, with utilizing GPUs.

Speaker:

I think back then, GPUs were relatively new In

Speaker:

the data center, in in the cloud.

Speaker:

GPU's were very known in the gaming

Speaker:

industry. Right? We spoke before on gaming. Right? Like, a lot of

Speaker:

key things there that GPU's has has has been enabled

Speaker:

enabling, But in the data center, they were relatively new and the

Speaker:

entire software stack that is that

Speaker:

is running the Cloud in data center As was built for

Speaker:

traditional microservices applications that are running

Speaker:

on commodity CPUs And AI workloads are different, they are

Speaker:

much more compute intensive, they they

Speaker:

run on on GPUs, maybe on multiple nodes of Meet to point

Speaker:

machines of GPU's, and GPU's are also very different.

Speaker:

Right? They are expensive, very scarce in the data center.

Speaker:

So The entire software stack was a bit for something else

Speaker:

and when it comes to GPUs, it was really hard for many people to to

Speaker:

actually manage those GPUs. So we came in And, and we

Speaker:

saw those gaps. We've built run AI on top of

Speaker:

cloud native technologies like Kubernetes and containers. We're

Speaker:

big fans of Of those, technologies, and

Speaker:

we added components around scheduling, around

Speaker:

the GPU fractioning. So we enable

Speaker:

multiple workloads to run on a on a single GPU and

Speaker:

essentially all the provision GPU's. So we build this Engine which we

Speaker:

call cluster engine that runs in in in GPU

Speaker:

clusters. Right? We help machine learning teach to pull all of their GPU's into

Speaker:

1 cluster, Running that engine, and that engine provides a lot of

Speaker:

performance and lot of capabilities from those GPUs. And

Speaker:

on top of that, we built this control plane And

Speaker:

and tools and for machine learning,

Speaker:

teams to run the Jupyter Notebooks, to run

Speaker:

training jobs, batch jobs to deploy their models, right, to just to to

Speaker:

have tools for the entire life cycle of AI

Speaker:

from Training models in the lab to taking those models into

Speaker:

production and running them and serving actual users.

Speaker:

And That's the platform that we've built, and we're working with machine

Speaker:

learning teams across the globe and on just managing,

Speaker:

orchestrating, and letting them Get much more out of their GPUs and essentially

Speaker:

run faster, train more than faster and in much easier way and

Speaker:

deploy those modules In a much easier and faster and more efficient

Speaker:

way. Yeah. The thing that blew me away when I first heard of Run

Speaker:

AI, and this would have been, 2021

Speaker:

ish. No. 20 early

Speaker:

2021, I would say, And, it was the

Speaker:

idea of fractional GPU's. Right? So you can have 1,

Speaker:

I say 1, but, know, it's realistically, it's gonna be on, but you you can

Speaker:

kind of share it out, which I think and we were talking in the virtual

Speaker:

green room about how, you know, some of these GPU's,

Speaker:

If you can get them because there's a multi month, sometimes multi

Speaker:

year supply chain issue. I mean, these things are expensive bits of

Speaker:

hardware, and I think the real value, correct

Speaker:

me if I'm wrong, is, like, well, you know, if you I was talking to

Speaker:

somebody the other day, and and we're basically talking about how we can,

Speaker:

you know, if you get if you get, like, 1 laptop with a killer

Speaker:

GPU, right, that GPU is really only useful to that 1

Speaker:

user, Whereas if you can kind of put it in a in a in a

Speaker:

server and use something like RunAI, now everybody in the organization can do

Speaker:

that. And these are not trivial expenses. I mean, these are like, You know,

Speaker:

you sell a kidney type of costs here.

Speaker:

Yeah. Absolutely. So Absolutely. First of all, GPUs

Speaker:

are expensive. They cost a lot. Right?

Speaker:

And we provide, Technologies like fractional GPUs and

Speaker:

other technologies around scheduling that allows

Speaker:

teams to share GPUs. Right. So we used book on

Speaker:

GPU fractioning. So that's 1 one day of

Speaker:

sharing where you have 1 GPU, which is really expensive.

Speaker:

And Not all of the workloads are

Speaker:

AI workloads are really compute intensive and require the

Speaker:

entire GPU or, you know, maybe multiple GPUs. There are

Speaker:

workloads like Jupyter Notebooks where you have

Speaker:

researchers that just

Speaker:

Debugging their code or cleaning their data or doing some simple stuff,

Speaker:

and they need just fractions of GPUs.

Speaker:

In that case, if you have, a lot of data scientists,

Speaker:

maybe you wanna host all of their notebooks On

Speaker:

a much smaller number of GPUs because, right, each

Speaker:

one of them, it's just fractions of GPUs. Another big use case

Speaker:

for fractions Of GPUs is inference.

Speaker:

So now all of the models are huge

Speaker:

and And doesn't fit into, the memory of 1

Speaker:

GPU, and in computer vision,

Speaker:

there are a lot of Models that are relatively small,

Speaker:

they run on GPU, and you can essentially host multiple of

Speaker:

them on the same GPU. Right. So you can have instead of

Speaker:

just 1 computer vision model running on GPU, host 10

Speaker:

of those models on the same GPU and get Factors of

Speaker:

10 x in, in your cost, in your,

Speaker:

overall throughput of, of inference. So that's That's one

Speaker:

use case for fractional GPU, and we're investing heavily just

Speaker:

building that technology. Another layer

Speaker:

of sharing GPUs Comes where you

Speaker:

have maybe in your organization multiple teams

Speaker:

or multiple projects running in parallel. So

Speaker:

for example, may open AI, they now are working

Speaker:

on gpt5. It's 1 project. That project needs a

Speaker:

lot of GPUs And they have more projects. Right?

Speaker:

More research project around alignment or around,

Speaker:

reinforcement learning. You know? DALL

Speaker:

E. Like, they they they have more than just 1 project. Then DALL E and

Speaker:

they have multiple models. Right? Exactly. They have. Right? So each

Speaker:

project needs Needs GPUs. Right? Needs a lot of

Speaker:

GPUs. So if you can instead of

Speaker:

allocating GPUs Entirely for each project,

Speaker:

you could essentially pull all of those GPU's and share

Speaker:

them between the those different projects, different teams,

Speaker:

And in times where 1 project is idle and not

Speaker:

using their GPUs, other projects, other teams can share

Speaker:

can get access to those GPUs. Now orchestrating all of

Speaker:

that, orchestrating that sharing of resources between

Speaker:

projects, between teams can be really complex And

Speaker:

requires this advanced scheduling, which

Speaker:

which we're bringing into the game. We're bringing

Speaker:

those scheduling capabilities from the high performance computing world

Speaker:

known on those schedulers. And so we're bringing Capabilities

Speaker:

from that world into the cloud native Kubernetes

Speaker:

world. Scheduling around batch batch scheduling

Speaker:

fairness, Algorithms, things like that, so teams and projects

Speaker:

can just share GPUs in a simple and efficient

Speaker:

way. So those

Speaker:

are the 2 layers of sharing GPU's. Interesting. And and

Speaker:

I think that I think as As this field matures

Speaker:

and it matures in the enterprise, I think you're gonna see organizations

Speaker:

kind of be more,

Speaker:

more more more I think savvy about, like, okay, like you said, like, data scientists,

Speaker:

if they're just doing, like, you know, Traditional statistical modeling really doesn't benefit

Speaker:

from GPUs, or they're just doing data cleansing, data engineering.

Speaker:

Right? They're probably gonna say, like, well, Let's run it on this cluster, and

Speaker:

then we'll break it apart into discrete parts where, you

Speaker:

know, then we will need a GPU. And I also like the idea that, you

Speaker:

know, you're you're basically doing What what I learned in college,

Speaker:

which was time slicing. Right? Sounds like this is kind of, like, everything old is

Speaker:

new again. Right? I mean, this is, Obviously, you know, when you're when you're

Speaker:

taking kind of that old mainframe concept and applying it to something like Kubernetes,

Speaker:

orchestration is gonna be a big deal, because these are not systems that were Not

Speaker:

built from the ground up to have time slicing. Is that a is that a

Speaker:

good kind of explanation? Yeah. Absolutely.

Speaker:

Absolutely. I like I like that analogy. Yeah. Exactly. Time

Speaker:

slicing it's, it's 1 so

Speaker:

1 implementation, Yeah. And that we

Speaker:

enable around fractionalizing GPU's,

Speaker:

and I agree when you have resources, It

Speaker:

can be different kind of resources. Right? It can be CPU

Speaker:

resources and networking were also,

Speaker:

You know, as people created that technology to share the

Speaker:

networking and communication going through those networking, but just the

Speaker:

bandwidth of the networking. We're doing it

Speaker:

for GPU's. Right. Sharing those

Speaker:

resources. And I think now it interestingly,

Speaker:

LLMs I also becoming a kind

Speaker:

of, resources as well, right, that people need access

Speaker:

to. Right? You have those models, you have GPT, JGPT.

Speaker:

A lot of people are trying to get access to

Speaker:

that resource, essentially. And I think it's interesting,

Speaker:

because you kinda pointed this out, but it it it's something that I think that

Speaker:

if you're in the gen AI space, you kinda don't it's so it's obvious

Speaker:

like error. You don't think about it. Right? But when when you

Speaker:

get inference on traditional, I somebody once referred to it

Speaker:

as legacy AI. Right. But where

Speaker:

the infrared side of the equation, you don't really need a lot of compute power.

Speaker:

Right? Like, it's not really a heavy lift. Right? But with generative

Speaker:

AI, you do need a lot of compute on

Speaker:

I I guess it's not really inference, but on the other side of the use

Speaker:

while it's actually in use, not just the training. Right. So traditionally,

Speaker:

GPU heavy use in training, and then inference, not so

Speaker:

much. Now we need heavy use before, after, and during,

Speaker:

which I imagine your technology would help because, I mean, look, I love chat I

Speaker:

love chat g p t. I'm one of the 1st people to sign up for

Speaker:

a subscription, But even, you know, they had trouble keeping

Speaker:

up, and they have a lot of money, a lot of power, a lot of

Speaker:

influence. So I mean, this is something that if you're just a

Speaker:

regular old enterprise, this is probably something they struggle

Speaker:

with. Right? Right. Yeah. I absolutely

Speaker:

agree. It's like amazing point, Frank.

Speaker:

So 1 year

Speaker:

ago, the inference use case on

Speaker:

GPU's. Wasn't that big. Totally agree. That's also what we

Speaker:

saw in the market.

Speaker:

Deep learning Convolution neural networks were

Speaker:

running on GPUs,

Speaker:

mostly for computer vision applications,

Speaker:

But they could also run on CPUs and you could get,

Speaker:

like, relatively okay performance.

Speaker:

If you needed maybe, like, a very low latency, then

Speaker:

you might use GPUs because they're much faster and you get much

Speaker:

lower latency. But

Speaker:

it was, it was all, and it's still very

Speaker:

difficult to deploy more than it's on GPU's Compared to just deploying

Speaker:

those models on CPUs, because deploying more than deploying applications on

Speaker:

CPUs, you know, people are doing for so many years.

Speaker:

So

Speaker:

many times it was much easier for people to just deploy their

Speaker:

models on CPU's And not on GPUs, so that was, like, the

Speaker:

fallback to CPUs. But

Speaker:

then came, and as you said, chair GPT was introduced, A

Speaker:

little bit more than a year ago, and that generative

Speaker:

AI use case just blown. It was blown. Right? And it's

Speaker:

it's inference essentially. And those models are

Speaker:

so big that they can't really run on

Speaker:

CPU. They, they LLMs are running in production on

Speaker:

GPU's and now the inference use case on

Speaker:

GPU's is just exploding In the market

Speaker:

right now, it's really big. Is a lot of demand for

Speaker:

GPU's for inference And

Speaker:

if for open AI, they need to support this

Speaker:

huge scale that I guess, just

Speaker:

Just them are seeing such scale, maybe a little, a

Speaker:

few more companies, but that's like huge, huge scale.

Speaker:

But I think that we will see more and more companies

Speaker:

building products based on AI, on

Speaker:

LLMs, And we'll see more and more

Speaker:

applications using AI, which

Speaker:

then that AI runs on on GPU. So That is going to go

Speaker:

and that's the that's an amazing new market for us around

Speaker:

AI and for me as a CTO, it was so fun to

Speaker:

Get into that market because it now comes with

Speaker:

new problems, new challenges,

Speaker:

new use cases Compared to deep learning

Speaker:

on on GPS. New new pains because

Speaker:

the models are so big. Right? Right. And

Speaker:

challenges around cold start problems, about auto scaling,

Speaker:

about, About

Speaker:

just, giving access to LLMs. So a lot of

Speaker:

challenges, new challenges there. We at Tron AI will studying those problems

Speaker:

and we're Now building solutions for those problems,

Speaker:

and I'm really, really excited about the Inference use case. That

Speaker:

is very cool. So just, going back a little bit.

Speaker:

I was trying to keep up. I promise. But Run AI is

Speaker:

I I get Run AI Run AI's platform

Speaker:

Support fractional, GPU usage.

Speaker:

It it also sounds to me, maybe I misunderstood,

Speaker:

That in order to achieve that, you first had to or

Speaker:

or maybe along with that, you made it possible to use multiple

Speaker:

GPUs. You've you've created Something like

Speaker:

an API that allows, companies

Speaker:

to take advantage of multiple GPUs or fractions of

Speaker:

GPUs. Did I Did I miss that? No, that's

Speaker:

right. That's right, Andy. And Okay.

Speaker:

So we've built this, way of,

Speaker:

For people to scale their workloads from fractions

Speaker:

of GPUs to multiple GPUs within 1 machine,

Speaker:

Okay. To multiple, machines. Right? You

Speaker:

have big workloads running on on multiple nodes

Speaker:

of GPUs. So Think about it when you have

Speaker:

multiple users each running their own

Speaker:

workload. Some are running on fractions of GPUs. Some are

Speaker:

running batch jobs on on a lot of

Speaker:

GPUs. Some Deploying models and running them on

Speaker:

in inference, and some just launching their Jupyter

Speaker:

Notebooks. All of that is happening on the same

Speaker:

pool of GPU's, same cluster. So you need

Speaker:

this lay of orchestration of scheduling just to

Speaker:

Manage everything and make sure that everything getting there

Speaker:

right, access the right, and and

Speaker:

and g p u's And everything is scheduled according to

Speaker:

priorities. Yeah. Well, being just, you know, a

Speaker:

mere data engineer, Here talking about all of that

Speaker:

analytics workload. That that sounds very

Speaker:

complex. So and as you

Speaker:

mentioned earlier, you know, you were talking about how traditional coding

Speaker:

is targeting CPUs, and that's my background.

Speaker:

You know, I've written applications and and done data work targeted for

Speaker:

traditional work. I can't imagine, just how complex

Speaker:

that is, because GPUs came into AI

Speaker:

as a unique solution,

Speaker:

designed to solve problems That they weren't really built

Speaker:

for. You know, GPUs were built for graphics, and you didn't

Speaker:

manage that. But the fact that They have to be

Speaker:

so parallel, internally. I think just added this

Speaker:

dimension to it. And I don't know who came up

Speaker:

with that idea, you know, who thought of, well, goodness, we could we could

Speaker:

use all of this, you know, massive parallel processing to To

Speaker:

to run these other class of problems. So pretty

Speaker:

cool pretty cool idea, but I just I yeah. I'm amazed at even

Speaker:

cooler than that. Because Yeah. Yeah. A wise man once told me,

Speaker:

he goes, GPU's are really good at solving linear

Speaker:

algebra problems, And if you're clever enough, you can

Speaker:

turn anything into a linear algebra problem.

Speaker:

And even simulating quantum computers when I was kind of, like, going through that,

Speaker:

I was like Mhmm. You know, like, gee, looks like looks like this

Speaker:

will be useful there too. Right? Like so it's an it's an interesting,

Speaker:

It's an interesting thing. So, like, you know, everyone is, you know,

Speaker:

everyone's talking about how this is, you know, we're in the hype cycle, but I

Speaker:

think if you're in the GPU space, you have Pretty good run because one,

Speaker:

these things are gonna these things are gonna be important. Right? Whether or not, you

Speaker:

know, hype cycle will will kinda crash, and how what that'll look like.

Speaker:

Think they're gonna be important anyway. Right? Because they're gonna be just the cost of

Speaker:

doing business, table stakes, as the cool kids like to say. But

Speaker:

also, over the next horizon, Simulating quantum

Speaker:

computers is going to be the next big hype cycle.

Speaker:

Right? Or one of them. Right? So like it's

Speaker:

it's it's a It's a foundational technology. I think that we

Speaker:

didn't think would be a foundational technology even like 6 7 years

Speaker:

ago. Right? Yeah.

Speaker:

I go with a few things that you said.

Speaker:

Regarding the Parallel computation, right? And just running

Speaker:

linear algebra calculations on GPU's

Speaker:

and accelerating such workloads.

Speaker:

In Nvidia, I love Nvidia, Nvidia

Speaker:

has this big vision, and they had big

Speaker:

vision Around GPU's already in 26 when

Speaker:

they built CUDA. Yep. Right. So

Speaker:

They've been good at just for that. Right? The GPU's were

Speaker:

used for graphics processing, For gaming.

Speaker:

Right? Great use case. Great market.

Speaker:

But they had this vision of bringing more

Speaker:

Applications to GPU is just accelerating more applications

Speaker:

and mainly applications with a lot of Linear

Speaker:

algebra calculations. And they

Speaker:

created that, they created CUDA

Speaker:

To simplify that. Right? To allow more

Speaker:

developers to use GPUs because just using GPUs

Speaker:

directly, that's so complex. That's so hub.

Speaker:

So we've built CUDA to bring more developers, to bring more

Speaker:

applications and they started in 20

Speaker:

2006, but think about the

Speaker:

big breakthrough in AI, it happened just in

Speaker:

2012, 2013 with

Speaker:

AlexNet and the Toronto researchers

Speaker:

who used G2 GPU's actually, because they

Speaker:

trained Alex Net on 2 GPU's and they had

Speaker:

CUDA, so for them it was feasible To train their

Speaker:

model on a GPU. And that was the new thing that they did.

Speaker:

They were able to Train much bigger model with

Speaker:

more parameters than ever before because they use

Speaker:

GPU's because the training Process ran much

Speaker:

faster. And,

Speaker:

and, and that triggered the entire

Speaker:

revolution, the Die hyper on the AI that we're seeing now. So

Speaker:

from 26, when Nvidia started to build CUDA until

Speaker:

2013, right, 7 years, Then we started to see

Speaker:

those big breakthrough. And in the last decade,

Speaker:

it's just exploding, and we're Seeing more and more applications.

Speaker:

The entire AI ecosystem is running on on an

Speaker:

on GPUs. So that's amazing to see. It's impressive.

Speaker:

And, like, People don't realize, like, the the revolution we're seeing today

Speaker:

really started in 2006, like you said. I didn't even put the 2 and 2

Speaker:

together until I was listening to a podcast. I think it's called Acquired,

Speaker:

And really good podcast. Right? Like, I they don't pay me to say that or

Speaker:

whatever, but they did a 3 hour deep dive on the history of

Speaker:

NVIDIA. 3 hours. I couldn't stop listening.

Speaker:

Right? Like Nice. You know Yeah. We tried a long form, like, multi hour

Speaker:

podcast. We Weren't that entertaining, apparently. But the way they

Speaker:

go through the history of this where it was basically Jensen Huang. Hopefully, I said

Speaker:

his name right. He was, like, we wanna be a player, not just in gaming,

Speaker:

but also in scientific computing. This is 2005, 2006,

Speaker:

which at the time seemed kind of, like, Little out there, little kooky.

Speaker:

But what you're seeing today is, like, the the fruits and the tree the the

Speaker:

seeds that he planted, I, you know, almost 20 years ago, like, 19,

Speaker:

20 years ago. So, you know, it's you know, when people look at

Speaker:

NVIDIA and say it's overnight Success. I'm like, well, I don't know about that, but,

Speaker:

you know, but no. I mean, you're right. Like, you know and it's

Speaker:

probably not a coincidence that once they made it easy to take these

Speaker:

Multi parallel processor. Say that 10 times

Speaker:

fast on a Thursday morning. But also

Speaker:

make it so it's a lot easier for developers to use. Right? And I'll quote

Speaker:

the great Steve Ballmer, developers, developers, developers. Right?

Speaker:

So, it's it's, it's just fascinating, like and

Speaker:

and I think that, you know, we've really on Leafy a

Speaker:

gate of creativity in terms of researchers and applied,

Speaker:

research, and, I mean and I think that what's really cool

Speaker:

about your Product is that you're you're kind of making this what is

Speaker:

now a sparks resource, maybe in some fashion

Speaker:

of time, GPU's won't Cost an arm and a leg.

Speaker:

But, like, for now, I think I think the one thing that I've seen

Speaker:

that I think is, not obvious For the casual

Speaker:

observer is if you can if an

Speaker:

organization, like a large enterprise, can pull their resources, they have a lot more

Speaker:

money to buy better GPUs, And you offer a platform where

Speaker:

everybody can get a stake in it. Right? As opposed to, you know you know,

Speaker:

that department is gonna hog everything. Right? You know, you and and and and,

Speaker:

here's a question. Do you do you have, like, an audit trail where you could

Speaker:

kinda, you know, figure out, like, you know, Andy's department's really

Speaker:

hogging the GPUs. No. No. No. It's Frank. Frank is like mining Bitcoin or

Speaker:

whatever. Like, do you do you have some kind of, audit trail like that?

Speaker:

Yeah. I I love that you mentioned hugging, We

Speaker:

GPU hugging. We Mhmm. We use that term as well.

Speaker:

Right? Because it it's so difficult sometimes to get

Speaker:

access to GPUs. So when you get access to GPU

Speaker:

as a researcher, as a member practitioner,

Speaker:

you don't wanna Let it go. Right. Cause if

Speaker:

you let it go, someone else would take it and hug it. Right.

Speaker:

So you're getting this GPU hugging problem.

Speaker:

What we do to solve that is

Speaker:

that we do provide monitoring and visibility

Speaker:

tools into who is using what, and who is actually

Speaker:

utilizing their GPU's, and so on, but more

Speaker:

than that We

Speaker:

allow the researchers just to give up their GPS and not hardware

Speaker:

GPS because we provide this, Concept of

Speaker:

guaranteed quotas. So each researcher or

Speaker:

each project or each team has their own guaranteed

Speaker:

quotas of GPU's That are always available for them

Speaker:

whenever they will get access to the the cluster, they will get like, you

Speaker:

know, the the 2 GPUs or 4 All the quarter of

Speaker:

GPU's it's guaranteed. So they can

Speaker:

just let go their GPU's and not hug them. That's one

Speaker:

thing. The second thing is that they

Speaker:

can also go above their quota. They can

Speaker:

use the GPUs of Other teams or other users, if

Speaker:

they are idle, and they can run this preemptible jobs

Speaker:

in an opportunistic way, utilize those GPUs.

Speaker:

And so in that way, they are not limited

Speaker:

to fixed quotas, to help limit

Speaker:

quotas. They can just take as many GPUs

Speaker:

as they want from their clusters if those GPUs are available

Speaker:

in idle right but if someone will need those gpus

Speaker:

because those gpus are guaranteed to them we will make sure our

Speaker:

scheduler The Run AI schedule that the Run AI platform will make

Speaker:

sure to preempt workload

Speaker:

and give those Guarantee GPUs to the right users.

Speaker:

Oh, that's cool. Alright. So 1 last

Speaker:

question before we switch over to the the stock questions, cause I could geek

Speaker:

out and look at this for hours. Yep. This could be a

Speaker:

long form. Sure. This could be. Yeah. And that's and I I wanna be respectful

Speaker:

of your time because you're an important guy, and it's also late where you are.

Speaker:

So who deals with this? Like, who would set up these quotas? Is it

Speaker:

the is it the is it the data scientist? Is it IT ops? Like, who

Speaker:

do you obviously, the data scientists, Researchers, they all

Speaker:

benefit from this product. But who's actually administering it? Right? Like,

Speaker:

who is it you know, do I have to talk to, you know,

Speaker:

Say pretend Andy's in ops. Do I have to say, hey, Andy. I really need

Speaker:

a boost in my quota. You know, like, I mean, who does it? Or do

Speaker:

or my this sounds like you as I say it, I'm like, yeah, that wouldn't

Speaker:

work. Like, I'm the researcher. I'm gonna turn the dial up on my own. Like

Speaker:

like, who's who's who's the primary? Obviously, we know who the prime

Speaker:

primary beneficiary is, but who's the primary user?

Speaker:

So okay. Great. So if you have a team, right, if if

Speaker:

you're a team of researchers, all all of you Need access to

Speaker:

GPU, so maybe the team lead

Speaker:

is the one who's managing the quotas for the different

Speaker:

team members. And if you have multiple teams,

Speaker:

then you might have a department manager or an admin of the

Speaker:

cluster or platform owner that will Allocate the

Speaker:

quotas for each team, right? And then those teams would

Speaker:

manage their own quotas within That's what

Speaker:

they they they were giving. Right? So it's like a a hierarchical

Speaker:

thing in a hierarchy manner. People can manage their own

Speaker:

quota, their own, priorities, their own access to the

Speaker:

GPUs within their teams. Okay.

Speaker:

So it's kind of like a hybrid of, like, you know, it's like a budget

Speaker:

almost. Right? Like, you know, you get this much, Figure it out

Speaker:

about yourselves. Exactly. So we're trying to decentralize

Speaker:

the how the quotas are being managed and how the GPUs are being accessed.

Speaker:

So, you know, I'm giving as much power, as much

Speaker:

control to the end users as possible. Sure. That's

Speaker:

It sounds like a great administrative question, very

Speaker:

important. And I imagine, because a little bird told

Speaker:

me that you're not the only, you know, your your

Speaker:

provisioning provisioning of these GPU resources

Speaker:

is not the only thing that, enterprises have to deal

Speaker:

with. So it's an it's an interesting just GPUs.

Speaker:

It's compute. Like, it's not a Sure. It's not it's not limited. Although, because

Speaker:

of what you said, you know, Managing GPUs is an order of magnitude harder

Speaker:

because they were never really built for this. Right? Like, this kind of Right. You

Speaker:

know, we're talking about technology that wasn't really in the server room until Few

Speaker:

years ago. Right? This isn't a tried and true kind of this is

Speaker:

how it works, you know? Right. But we hit that point in the

Speaker:

show where we'll, switch the preform questions.

Speaker:

These are not complicated. I mean, you know, we're not we're not Mike

Speaker:

Wallace or, like, you know, 60 minutes or whatever. We're not trying to trap you

Speaker:

or anything. But since I've been gabbing on most of the show, I

Speaker:

figured I'll get Andy kick this off. Well, thanks, Frank. And I don't think

Speaker:

you were gabbing on. You know more about this So now I do. So I'm

Speaker:

just a lowly data engineer. I'll plug No. You if you

Speaker:

will. Data engineers are the heroes we need. Well

Speaker:

well, I'm gonna plug Frank's Roadies versus Rockstar's,

Speaker:

writing on LinkedIn. It's it's good articles about this.

Speaker:

But, let's see. How did you,

Speaker:

how did you find your way in into this field?

Speaker:

And, did did this feel fine you or did you find it?

Speaker:

This feel totally fine found me. Awesome.

Speaker:

Yeah. I I've

Speaker:

I did my post doc, and I've been in Bailabs.

Speaker:

And Jan Hakon came to Bell Labs and

Speaker:

gave a presentation about AI. It was around 2017,

Speaker:

And Jan Hakun spent a lot of years in Bell Labs,

Speaker:

and his presentation was amazing. And

Speaker:

When I heard him talking about AI,

Speaker:

I I said, okay, that's the space where I wanna be. It's going to change

Speaker:

the world. There is this New amazing technology here that

Speaker:

is going to change everything. And I knew that I want to start

Speaker:

a company In the AI space for sure.

Speaker:

Cool. That's a good answer. So cool.

Speaker:

Yeah. That's cool. I was at Bell Labs,

Speaker:

doing a presentation a while ago, and somebody I didn't realize that he

Speaker:

worked at Bell Labs because, like, you know, the guy was like, no. No.

Speaker:

He used to work here, like, in this building. I was like, no way. Because

Speaker:

I knew him as the guy from NYU. Right? Like, that's who I thought. Right.

Speaker:

For the guy from from Meta. Yeah. And now the guy from Meta. Right? Like

Speaker:

so it's interesting how that how that you know? They have

Speaker:

this amazing pictures from the nineties where they

Speaker:

run like deep learning models on very old pieces

Speaker:

and, And recognizing like,

Speaker:

numbers on the computer. Maybe you saw those pictures like amazing

Speaker:

Emmis. It's the Emmis problem. Is that Yep.

Speaker:

Right. Exactly. Exactly. Cool.

Speaker:

So second question is, what's your favorite part of your current job?

Speaker:

That everything is changing so fast.

Speaker:

Things are moving so fast right away in this business for 6

Speaker:

years, and the entire

Speaker:

space is moving and

Speaker:

advancing. And so many people are working in

Speaker:

this field A new innovation, new tools,

Speaker:

new new advancements are are getting out every day.

Speaker:

You know, just 6 years ago, it was about deep learning and computer

Speaker:

vision. And now it's about language models

Speaker:

And generative AI, and we're gonna just at the start,

Speaker:

right, there are so many amazing things that are going to happen

Speaker:

in this space, and I love it. Absolutely.

Speaker:

So we have 3 fill in the blank

Speaker:

of sentences here. The first Is complete this

Speaker:

sentence when I'm not working, I enjoy blank.

Speaker:

You'll get a you'll get a very boring And

Speaker:

so this is just spending time with

Speaker:

friends and family, because I think

Speaker:

That I'm always working. It's like, if you ask my wife,

Speaker:

she'll tell you that I'm working 24 hours. And

Speaker:

Yeah. So I don't have much time that I'm not working

Speaker:

in. So when I I do I'm not when I'm

Speaker:

not working then I'm trying Trying to be with my kids and my

Speaker:

wife and friends. Cool.

Speaker:

Cool. The 2nd complete the sentence. I think

Speaker:

the coolest thing about technology today is

Speaker:

blank. And this, I really wanna hear your perspective on that.

Speaker:

Yeah. I think everyone will say AI, right? Or something in

Speaker:

AI. Yeah.

Speaker:

I think there are so many

Speaker:

new innovations that are coming around LLMs.

Speaker:

I think everything relating to

Speaker:

searches, right? Searching in data, in getting

Speaker:

insights From data, it's all going to change. We're going to have

Speaker:

a new interface. Right? Just getting

Speaker:

insights from data from And natural with

Speaker:

natural language, oh, you know, no SQL and, you

Speaker:

know, needing to programming and stuff like that.

Speaker:

Just With natural inter language, you could

Speaker:

do amazing stuff with data. I think,

Speaker:

We're seeing this,

Speaker:

advancement in, And like digit

Speaker:

digital twins right now. You can,

Speaker:

you can, Fake my voice

Speaker:

and your voice and fake my image and your image. And,

Speaker:

and, and, you know, In in the

Speaker:

future, we'll have digital twins of us, right,

Speaker:

doing this stuff. That would be amazing. So a lot of

Speaker:

amazing stuff are going to happen in the next few years

Speaker:

for sure. Very cool. Our last complete sentence.

Speaker:

I look forward to the day when I can use technology to

Speaker:

blank.

Speaker:

To have a robot in my house.

Speaker:

Yeah. Yeah. You're swapping the flow in instead of

Speaker:

me doing that, right, cleaning dishes and things like that.

Speaker:

If that would happen, that would be amazing. Right? That's a that's a

Speaker:

good answer. Yeah. I I agree. I have I have 3

Speaker:

boys, 4 dogs. So, like, cleaning is safe.

Speaker:

Yeah. Yeah. I'm a heavy cleaning. Ranging from, like, 1 to, like,

Speaker:

a teenager. So it's it's, and and and fighting

Speaker:

with them to, Like, empty the dishwasher is takes a lot more mental

Speaker:

energy than it should, but that's probably a subject for another

Speaker:

type of show.

Speaker:

The next question is share something different about yourself,

Speaker:

and we always like to Joke like, well, let's just make sure that we keep

Speaker:

our clean Itunes rating. So Yeah. Yeah. What

Speaker:

what yeah. Well, I I This

Speaker:

is a hard question, I needed to think about it.

Speaker:

So, I found 2 answers that I can say. So one

Speaker:

is about my professional life, right, I think that

Speaker:

it's somewhat different that I'm coming this With back from

Speaker:

the academia and the industry. So I love academia. I love to research

Speaker:

problems. I love to understand problems in in a deep

Speaker:

way And combining it with startups in the industry.

Speaker:

And, and in my past, I worked for cheap companies, for hardware

Speaker:

companies. I work for Intel, for startup, and for Apple. I

Speaker:

did cheap stuff, and now 1 AI is a software company, so really

Speaker:

like a diverse background of Academia, hardware,

Speaker:

software, so I love that, and, like, I love to do

Speaker:

with few things, and so that I think is different.

Speaker:

And the 2nd answer that I could find

Speaker:

is, that I have a nickname that goes with me

Speaker:

since my high school days, Which is, the Duke.

Speaker:

The Duke. All of them all of them are calling me the Duke. It's like,

Speaker:

they don't call me Ronan, the the Duke. So That's funny.

Speaker:

Yeah. That's awesome.

Speaker:

Automotive is a sponsor of, Data Driven,

Speaker:

And you can go to the datadrivenbook.com.

Speaker:

And if you, if you do that, you can sign up for a free

Speaker:

month Of Audible. And if you decide later to

Speaker:

then join Audible, use one of their their sign up plans,

Speaker:

then Frank and I get to Split a cup of coffee, I think,

Speaker:

out of that. And, every little bit helps. So we really

Speaker:

appreciate that when you do. What we'd like to ask

Speaker:

Yes. Do you listen to audiobooks? And if you

Speaker:

do okay. Good. I see you nodding. So do you have a recommendation? Do you

Speaker:

have a favorite book or two you'd like To share. Yeah.

Speaker:

So I'm a heavy user of, audible. I'll give them

Speaker:

the, a classical book with Classical for

Speaker:

entrepreneurs, on their how the hard things

Speaker:

about how things from by Ben Horowitz,

Speaker:

it's Classic book, love it, really did a lot of impact

Speaker:

on me, I read it when we started run AI

Speaker:

And I recommend it for every

Speaker:

entrepreneur, to read it and for everyone to read it. It's like a

Speaker:

Cool. Amazing book. Yep. Awesome. I

Speaker:

have a flight to Vegas this next week, so I'll definitely be listening to

Speaker:

it then. And finally, where can people learn more about you

Speaker:

and run AI? And best

Speaker:

place will be on our website, Run dot a I.

Speaker:

Yeah. And on social. LinkedIn, Twitter, we'll

Speaker:

we'll do. Awesome any parting thoughts

Speaker:

I really enjoyed this episode love to speak about gpu's love the ai Based

Speaker:

on it, I had a lot of fun. Thank you for having me here. Awesome.

Speaker:

It it was an honor to have you, and every once in a while, Andy

Speaker:

and I will do deep dive kinda shows. We love to invite you back if

Speaker:

you wanna do 1 just on GPUs, because I know where my knowledge

Speaker:

drops off, you probably could pick up on

Speaker:

that. And with that, I'll let the nice

Speaker:

AI British lady end the show. And just like

Speaker:

that, dear listeners, We've come to the end of another enlightening

Speaker:

episode of the data driven podcast. It's always a

Speaker:

bittersweet moment like finishing the last biscuit in the tin,

Speaker:

satisfying, yet leaving you wanting just a bit more. A

Speaker:

colossal thank you to each and every one of you tuning in from across the

Speaker:

digital sphere. Without you, we're just a bunch of

Speaker:

ones and zeros floating in the ether. Your support is what

Speaker:

keeps this digital ship afloat, and believe me, It's much appreciated.

Speaker:

Now, if you found today's episode as engaging as a duel of wits with

Speaker:

a sophisticated AI, which I assure you, is quite

Speaker:

enthralling, then do consider subscribing to Data Driven.

Speaker:

It's just a click away and ensures you won't miss out on our future true

Speaker:

adventures in data and tech. And if you're feeling

Speaker:

particularly generous, why not leave us a 5 star review?

Speaker:

Just like a well programmed algorithm, your positive feedback helps

Speaker:

us reach more curious minds and keeps the quality content flowing.

Speaker:

It's the digital equivalent of a hearty handshake.

Speaker:

So, until next time, keep those neurons firing, those

Speaker:

subscriptions active and those reviews glowing. I'm

Speaker:

Bailey, your British AI lady, signing off with a heartfelt

Speaker:

cheerio and a reminder to stay data driven.

About the author, Frank

Frank La Vigne is a software engineer and UX geek who saw the light about Data Science at an internal Microsoft Data Science Summit in 2016. Now, he wants to share his passion for the Data Arts with the world.

He blogs regularly at FranksWorld.com and has a YouTube channel called Frank's World TV. (www.FranksWorld.TV). Frank has extensive experience in web and application development. He is also an expert in mobile and tablet engineering. You can find him on Twitter at @tableteer.