Navigating the Complexity of Operationalizing ML Models
In this episode of Data Driven, our Andy Leonard and Frank La Vigne are joined by Chris McDermott, VP of Engineering at Wallaroo.AI. Together, they explore the challenges and advancements in the ever-evolving world of machine learning and artificial intelligence.
From the importance of ongoing care for machine learning models to the rise of edge computing and decentralized networks, they touch on the critical need for flexibility and data privacy. Chris shares his insights on the technical challenges of AI and ML adoption, as well as his unique career journey. They also discuss the evolution of technology and the potential future impact of these innovations.
Join us for a deep dive into the world of AI, technology, and the future of machine learning with Chris McDermott on this episode of Data Driven.
Show Notes
00:00 Exploring AI, data science, and data engineering.
06:20 Training and inferring are different stages.
08:12 Legacy AI doesn’t require neural networks or GPUs.
12:09 Machine learning models require consistent care and monitoring.
15:10 MLOps merges skills, breaks down silos, collaborates.
16:47 Prefer MLOps to avoid namespace collision. DevOps parallels original Star Wars plot.
20:27 Internet-scale operations require automation and resilience.
24:13 Challenges of integrating AI into business processes.
28:03 New push for edge computing in technology industry.
32:05 Edge technology critical, discussed in government tech symposium.
34:50 Navigating from SendGrid to Twilio simplified processes.
36:15 First foray into data, growing knowledge.
39:33 Technology evolves, builds complexity over time.
44:41 Book recommendation: “Seeing Like a State” by James C. Scott discusses legibility and centralization of power in society.
46:28 Predictable tree farming fails due to ecosystem complexity.
Speaker Bio
Chris McDermott is a software engineer and entrepreneur who is passionate about creating products that make machine learning more accessible and manageable for users. His focus is on developing a platform that allows for easy deployment and management of machine learning models using any framework and on any architecture or hardware. He believes that current solutions in the market force users into a specific platform, and he aims to provide a more flexible and efficient alternative. With a strong belief in the potential of his product, Chris is dedicated to making machine learning more accessible and user-friendly for people across various industries.
Transcript
Welcome to show 350 of data driven.
Speaker:In this episode, Frank and Andy interview Chris McDermott,
Speaker:VP of engineering at Wallaroo. Wallaroo helps
Speaker:customers operationalize machine learning to ROI in the cloud,
Speaker:in decentralized networks, and at the edge. It's
Speaker:a fun conversation on MLOps and the future of intelligence systems
Speaker:and model management. Now on to the show.
Speaker:Hello, and welcome to data driven, the podcast where we explore the emergent
Speaker:fields of artificial intelligence, Data science and of course, data
Speaker:engineering, the fundamental thing that kind of underpins it all. And with
Speaker:me on this epic road trip down the information superhighway
Speaker:Is my favorite data engineer of all time, Andy Leonard. How's it going
Speaker:Andy? Good Frank, how are you? I'm doing alright. I'm doing
Speaker:alright. We just, we're chatting
Speaker:in the, virtual green room about some of the logistical challenges we had,
Speaker:with Microsoft Bookings and how Kind of like you can only have, you
Speaker:know, like, remember that the pick any 2 triangle, right? Good, fast, and
Speaker:cheap? Yep. Yep. Like, we can only have 2 things, 2 features of what we
Speaker:needed to do. Right. Alright.
Speaker:Despite logistical challenges, we are excited here to have,
Speaker:Chris McDermott who is the VP of engineering at Wallaroo,
Speaker:and, he is a passionate, and intellectually
Speaker:curious professional With excellent communication skills, he
Speaker:loves hard problems, then he must have definitely loved the
Speaker:process to get on the show, And,
Speaker:have yet to meet 1 he couldn't solve somehow. Maybe we should get you,
Speaker:Chris, to help us with our scheduling stuff. Really? You visit
Speaker:later? Yeah. So welcome to the
Speaker:show, Chris. Thank you. Thank you. It's great to be on. It's nice to meet
Speaker:you both. Well, likewise. Likewise. So you're coming to us from the,
Speaker:Mile High City. That's right. Awesome place. It's, I was
Speaker:there once, for internal Microsoft
Speaker:conference actually. Oh, nice. And beautiful town, like, it was
Speaker:just really cool. I think it was the 2nd biggest
Speaker:event that in Denver history was the Microsoft thing. Wow.
Speaker:And they they literally ran out of hotel rooms like it was.
Speaker:Oh, wow. It was pretty wild. Yeah. I think it was, just
Speaker:before one of the big parties had a convention there. And,
Speaker:they Oh, yeah. Yeah. Yeah. Yeah. I was so I'm actually
Speaker:slated to head back there next year for a Red Hat
Speaker:conference, so we'll see Let's see if the hotel situation has
Speaker:improved. I think it's improved a little bit. The city's been growing a lot. So,
Speaker:Yeah. Lots of government. Isn't Denver the place that has, like,
Speaker:the large bear up against the conference center that
Speaker:Yeah. Yeah. Yeah. Yeah. Yeah. That's exactly right. A giant blue bear appearing in the
Speaker:window of the conference center. Yeah. I was there. And,
Speaker:and and I remembered that. That was the That was the first thing I
Speaker:remember. It was, I was there in
Speaker:2007 for a Kind of a Microsoft conference. It was
Speaker:a, Professional Association for SQL
Speaker:Server. That's what it was called back then. And, I was
Speaker:actually the first one I spoke at. I've spoken at a bunch since then,
Speaker:but:Speaker:yeah. Like, I echo what Frank said, beautiful city
Speaker:and, just very picture picturesque. Yeah.
Speaker:Yeah. The weather in the mountains are beautiful. Mhmm. Yeah. And
Speaker:it's funny, like, you know, on the East Coast, we talk about mountains,
Speaker:but It's nothing like that. Like Right. Yeah. We quite the
Speaker:same. You would laugh at what we call mountains. Yeah. Right. But
Speaker:I remember a Robin Williams bit Where he said something like that
Speaker:people he admired the people in Denver because they got to
Speaker:Denver and they looked at the mountains and went, Well, I can't say what
Speaker:he said, but he had a kind of an Elon
Speaker:moment.
Speaker:There's so many of those. There's so many No more. We're stopping right here. We're
Speaker:not going over those mountains. So,
Speaker:You're VP of engineering at Wallaroo. So tell us a little bit about Wallaroo.
Speaker:Mhmm. Plus you're also ex data robot too. That that's interesting.
Speaker:Yes. Yep. Exadata robot. Yeah. So I've been working in the machine
Speaker:learning and AI space for, about 7 years now, I guess, or 6
Speaker:years. And, it's been really fun. You know, it's, it's a
Speaker:good time to be in the business. There's a lot of development
Speaker:happening, very fast pace of change, which I appreciate.
Speaker:And, you know, Wallaroo has been really great. Like,
Speaker:the team is fantastic, and the people are wonderful. And it it's a lot of
Speaker:fun working, with people that you enjoy hanging out with and and you respect
Speaker:and everything, that's that's very important to me. That's awesome. But I also just
Speaker:I think the product is awesome. It's really, I think,
Speaker:playing well in the market. Like, we are focusing on making it as easy as
Speaker:possible to deploy And manage machine learning models.
Speaker:And the focus is really on any model using any framework and being
Speaker:able to deploy onto sort of any architecture, any hardware,
Speaker:and being able to leverage GPUs if you need them or different kinds of CPUs,
Speaker:different acceleration libraries that people have tailored to the different
Speaker:architectures. And, honestly, there are not a lot
Speaker:of other solutions that tackle those 2 problems for people. Right.
Speaker:A lot of the other companies that we're competing with, they are trying to be
Speaker:like an end to end solution or, like, really force people into, you know, their
Speaker:platform. So you train on their platform, you deploy on their platform, you manage on
Speaker:their platform. But it's very limiting in terms of what you can bring on to
Speaker:the platform and and being able to, deploy on the different types of
Speaker:architectures and, platforms and things like that. So it's really
Speaker:exciting. It's fun. I think that's really important that you bring up
Speaker:the CPU solutions. As I've been tinkering,
Speaker:you know, the past couple of years with, you know, with the
Speaker:different, different platforms that are out there, it's
Speaker:That's definitely a smaller market, but maybe it's emerging now. I'm
Speaker:just not sure. Mhmm. Yeah. I wonder yeah. Sorry. Go ahead.
Speaker:Well, I was gonna say, you know, a lot of the time people conflate training
Speaker:and inferring, which is, you know, sort of the 2 different stages. Like,
Speaker:1st, you have to train a model, but then you use the model to make
Speaker:inferences, which, you know, it's really like asking the model to make a prediction
Speaker:or you give it some input and it gives you some output. And,
Speaker:they're they're very, very different tasks. And just because, you know,
Speaker:like, you may wanna use some hardware GPUs for training Doesn't
Speaker:necessarily mean mean that you need the GPUs when you are in production and
Speaker:you're asking it for predictions. A lot of the time, you know,
Speaker:The model is small enough that you really don't need to, but there's
Speaker:so much hype. It it's hard sometimes to separate the hype from the, you
Speaker:know, The real stuff and Yeah. Yeah. The hype the hype
Speaker:machine is real. I mean, like, it's and and and I I wanna get your
Speaker:thoughts on, you know, I mean, I love generative AI. I'm not
Speaker:knocking generative AI, but it feels like it's taken all the oxygen out of the
Speaker:room for All the other kinds of AI.
Speaker:Yeah. Yeah. Yeah. Because there are a lot of, you
Speaker:know, great models. I like XGBoost is a very standard one. It's
Speaker:been around for, you know, a long time, meaning at least for, you know, 5
Speaker:or 10 years now or something. But, that really honestly solves so
Speaker:many problems, and it's such a Small, easy model to deploy.
Speaker:I I wish people would focus more on on that kind of thing rather than
Speaker:hype. Right. No. That's a good point. And I think you
Speaker:bring up an interesting point because not all not
Speaker:all AI workloads are created equal. Right? Obviously, there's,
Speaker:I heard this term the other day and I had to spit my coffee out
Speaker:because it was just so funny. Legacy AI. Yeah.
Speaker:Yeah. There's generative AI now. There's legacy AI. That's
Speaker:crazy talk. You know? And I was just like,
Speaker:wow. But,
Speaker:you know, because, you know, legacy AI, basically,
Speaker:you're not using deep learning, you're not using neural networks,
Speaker:Generally, you don't get a good boost from GPU's.
Speaker:Correct. Right. And that's something that when when you tell that to
Speaker:Even tactical decision makers, they they they
Speaker:kinda look at you like, you know, what sorcery is that? Like, you know, because
Speaker:they'll they'll They'll say, like, oh, we don't have enough GPUs. There's no budget for
Speaker:GPUs. Like, what what what types of workloads are you running? And I
Speaker:tell them, it's like, well, it's not really a concern for you. Like, you don't
Speaker:need them. Yeah. And you see, you know,
Speaker:the the the the people who are doing the actual data science, they're like, yeah,
Speaker:duh, that's what we're trying to tell you. Yeah. But you see, like, the leaders
Speaker:of these teams are like, like, you know,
Speaker:it's, now Just for my own education,
Speaker:there wasn't there something called RAPIDS, and it was an
Speaker:acronym that let you use GPU's For
Speaker:certain types of like XGBoost, I think was one of them. Random
Speaker:forest there. I don't know. Oh. You See, it's funny because
Speaker:it was an it's an NVIDIA thing and obviously it only optimizes on. But,
Speaker:like, it was I remember Hearing about
Speaker:it in:Speaker:everything, and you haven't heard of it. Only,
Speaker:like, one ever per other person I met in the wild has ever heard of
Speaker:it, and he was at the same conference I was at where we heard about
Speaker:it. So I'm like, That's kind of unusual,
Speaker:but, you know, we gotta watch so
Speaker:fast, you know, and it's really hard to tell sometimes What
Speaker:what, which new developments are gonna end up being the future and which ones
Speaker:are gonna end up as dead ends? Right. You know, and even all
Speaker:the transformer stuff that that is powering GPT and and those similar types
Speaker:of models, I think that was originally written up in a
Speaker:white paper in, like,:Speaker:while, and nobody paid a whole lot of attention to it until OpenAI really
Speaker:ran with it. So yeah. Pension is all you
Speaker:need. I think that's was that the paper? Sounds right.
Speaker:Yeah. And then we're gonna go. Oh, sorry. Go ahead. Sorry,
Speaker:Andy. I cut you off your point. No. I I don't wanna go too far
Speaker:downstream before I say cred boost for using the phrase I don't
Speaker:know. Oh, nice. Somebody with your
Speaker:credentials, you know, saying I don't know. That's that's super
Speaker:cool. So cred Honestly, there's way too much to know. There's no
Speaker:way anyone person could know that. I I like to joke. I
Speaker:haven't checked my phone or, like, news Feed in like half an hour
Speaker:and I'm like woefully behind now. Yeah.
Speaker:But it feels that way like in the whole Oh, no. It does. Yeah. Especially
Speaker:it was especially interesting when the whole drama on OpenAI, and I
Speaker:don't wanna go down that rabbit hole too far. But when all of that soap
Speaker:opera kinda unfolded Yeah. Yep. It was kind of like,
Speaker:what's the latest? Like, is he back? Is he gone? Is he working at Microsoft?
Speaker:Like, he did work at Microsoft for like 10 minutes, and now he doesn't.
Speaker:Like, Yeah. You know, at
Speaker:at some some point down the middle of it, it's like, call me when this
Speaker:is over, and I'll deal with the, things yeah. I'll
Speaker:check-in again. But that's just the human
Speaker:side of it, let alone the let alone technology side of it.
Speaker:So Operationalization. I think that's gonna
Speaker:be the buzzword. Obviously, chatty b t and JennyIs, taking
Speaker:all the air out of there. And I think the next buzzword It's gonna be
Speaker:operationalization. 1, because it's kinda
Speaker:hard to say, and I'm not gonna lie, I've had to practice.
Speaker:But, it's something that I think companies and organizations
Speaker:that adopt AI, whether it's legacy AI
Speaker:Or generative AI. They're gonna have to realize, like, it's one thing to build
Speaker:the model, and then it becomes a, okay, now
Speaker:what? Yeah. Yeah. Well and models
Speaker:really are just like any other software. It's not something that you just
Speaker:write once and you, you know, Throw it out there, and it runs forever
Speaker:without being touched. Right? All of it requires care and feeding, and
Speaker:and machine learning models are no different. So, I think
Speaker:part of it is, you know, how do you deploy it? And then, you know,
Speaker:how do you keep that that deployment up to date, you know, getting critical
Speaker:patches and vulnerability fixes and things like that. But also
Speaker:how do you monitor the model and how it's performing and how it's performing
Speaker:relative to the real world, Because the world doesn't stand
Speaker:still right. So even if the model was trained on some data and it was
Speaker:98% accurate when it was trained, as the world shifts and
Speaker:and the situation around it shifts, that accuracy will
Speaker:almost certainly start to degrade over time. So You need to monitor that. You need
Speaker:to know when to retrain the model. And you have to be kind of
Speaker:keeping track of, new training data. Right? So the
Speaker:the the new environment that the model is operating in, you need to be recording
Speaker:all of the the inputs and also paying attention to the ground truth of, You
Speaker:know, what was the outcome of that prediction that the model made? Was it accurate
Speaker:or not, after the fact? And and correlating that back into your
Speaker:training data So you can retrain the models and, you know, keep them going
Speaker:over time. And that's just, you know, assuming you're gonna be using the same model
Speaker:forever. But as we just finished talking about new models coming out all the
Speaker:time, new approaches, new techniques. So, yeah, it really is
Speaker:is something you've gotta pay attention to. It's an
Speaker:extremely Yeah. It's an extremely dynamic space.
Speaker:Mhmm. I've heard this called
Speaker:MLOps for the longest time. Mhmm. Mhmm. But I've also heard a new term
Speaker:kinda pop up on the radar called AI ops Mhmm.
Speaker:For this. What do you call it? I
Speaker:generally call it MLOps. You know, one, I
Speaker:I sort of per like, AI and ML, there's an
Speaker:interesting, you know, difference there in in terms of who uses the different terms and
Speaker:when they use them. For me, AI is
Speaker:more of a general term that I use conversationally. And most of the time if
Speaker:I'm trying to be fairly technical and specific, I'll usually revert to ML,
Speaker:Because in fact, most of these things are machine learning. AI is a much more
Speaker:nebulous concept, and I I don't even think everybody agrees on on what AI is
Speaker:or What the threshold would be, you know, if you're
Speaker:doing statistical analysis, I think most people probably would not call that
Speaker:AI. But there are a lot of machine learning models that do work that way.
Speaker:And and that's definitely, like, part of the gradient. You know? I've
Speaker:noticed that too. Like, there it is a gradient too. Like, there's not a, like,
Speaker:a hard, like, You know, typically it depends on the audience. Right? If they're
Speaker:if they're BDMs, business decision makers, they're gonna use
Speaker:AI. Yeah. They're technically focused people. They tend to prefer the term
Speaker:ML. Yeah. That's also been my experience Interesting.
Speaker:Quite often. So I like MLOps because, one,
Speaker:it sort of grounds you a little bit more in that technical perspective. Mhmm.
Speaker:And, and it's sort of a like, To me, I think I came up
Speaker:through DevOps a lot of my, you know, first half of my career was DevOps
Speaker:and infrastructure and things like that. And, I
Speaker:think part of the appeal of the term MLOps is it taps into a
Speaker:lot of the DevOps, associations. Right? And
Speaker:Right. The concepts and the themes of DevOps, which is really about,
Speaker:merging different skill sets and breaking down silos and getting different teams to
Speaker:communicate with each other and And to collaborate more,
Speaker:being more dynamic. Not just, you
Speaker:know, putting software out there and and letting it run forever, but Keeping it up
Speaker:to date, monitoring it, recording the logs, you know, all of that kind of
Speaker:stuff, and and getting into a flow of continuous
Speaker:deployment, you You know, continuous integration, continuous testing, continuous
Speaker:deployment. And I think on the ML side, that's also where
Speaker:MLOps really shines and and is bringing those themes
Speaker:to the party, rather than a data scientist training a
Speaker:model, deploying it, and, you know, Throwing it over the
Speaker:wall to to, like, an operations team or something. It's
Speaker:getting all these different teams and skill sets to work together. It's
Speaker:building a continuous, you know, pipeline, with
Speaker:monitoring and and feedback loops and so on. So that's that's why
Speaker:I like MLOps. No. I like that too. So in order to prevent
Speaker:any hate mail come in or or but actually comments, AI
Speaker:ops is also used, I've heard, in,
Speaker:the telcos and network operators tend to have a term
Speaker:called AI ops, where they use AI to help operate their network. So that is
Speaker:Got it. It's it's a it's a namespace collision,
Speaker:which I've free further which I prefer MLOps for to avoid the namespace
Speaker:collision, plus all the reasons you said. You
Speaker:know, what's interesting is and I came from a software engineering
Speaker:background and, you know, and I'll be honest, I was not
Speaker:initially a big, believer in in DevOps, but
Speaker:kind of as time went on, I became a convert. But I think
Speaker:that, you know, when you look at how AI models, ML
Speaker:models, whatever, how they get operationalized.
Speaker:You look at it And I I often I often can tell
Speaker:who the fans of the new Star Wars movies are by using this analogy,
Speaker:because I'll say it's The:Speaker:Wars movie and the:Speaker:DevOps being kinda like the original, episode 4 And then the
Speaker:new, the the first new one, right? It's the same
Speaker:plot. I mean, the characters have changed, some things are
Speaker:different, But very effectively, it's the same plot. And, you
Speaker:know, some people will laugh like you did, and some people I can
Speaker:see will, Their their faces turn red. But,
Speaker:but I mean it's like it's it's the same plat plot. The names, the places
Speaker:have some have changed. But you're right. I mean, I think and there's a lot
Speaker:of lessons we can learn Yeah. In the ML community
Speaker:from the DevOps world. Right? Because, You know prior
Speaker:to DevOps, you know, the developers and operations had a
Speaker:very antagonistic relationship for the most part. I'm sure there's always
Speaker:exceptions. You know, I was I was joking that they would only meet,
Speaker:they only have to interact 3 times a year, and one of those was the
Speaker:holiday Christmas party. You know what I mean? And
Speaker:Yeah. But if you wanna deploy something in a far more agile
Speaker:way where they have to, you know, you put it In some extreme cases, every
Speaker:few hours, some new bit of code gets gets pushed up. That's obviously on
Speaker:on one fore end of the spectrum. But for the most part, you know, a
Speaker:couple times a month is not unreasonable. You have to automate that. You have to
Speaker:have processes in place. Yep. And I I see a day, and if that day
Speaker:has not already come, I would be surprised, That AI is gonna be the
Speaker:same thing or ML. Right? You're you're gonna have to get but to your point,
Speaker:right, this is a continuous process, You know? Yep. Yeah.
Speaker:We can't get away with, you know, you have this isolated team of data scientists.
Speaker:They they they kinda go off to their little area 51 type labs
Speaker:in secret. Right. I then come back with some model,
Speaker:and and I'm guilty of this too. I've done this. Right? Where I'm like, I
Speaker:built the model. I'm done. I did the math. I did the hard
Speaker:part. How do you get the play it? Not my problem. Not my
Speaker:problem. And it's funny, like Yeah. You know, I caught
Speaker:myself. Right. I caught myself doing that as I, you know, you
Speaker:know, doing that. Like recently, I had to I had to do a demo
Speaker:and I had to work on a kind of a It's basically a predictive
Speaker:maintenance type thing, and I took all this data, had the model, and I
Speaker:just said, here's the here's the link to the model, Have at it,
Speaker:pal. Mhmm. And then as I sent that, I was like, you know, I should
Speaker:probably be more involved in getting this on a race for it.
Speaker:Right. Yeah. Yeah. Yeah. Yeah. No. I think that's a big part of it.
Speaker:Another big part of it though is, scale, you know. And I think scale
Speaker:scaling of compute and, how How people were using compute and how
Speaker:much compute was required was a big part of what drove DevOps.
Speaker:You know, if you were a sysadmin responsible for a 100
Speaker:servers, That's, you know, challenging, but it's feasible. Like, you can do
Speaker:that. You can keep them all up to date. You can keep them all in
Speaker:sync with each other. Make sure they they all have the same patch levels and
Speaker:and so on. But you scale that up to a 1,000 servers?
Speaker:That gets a lot trickier. You try to go to a 100,000 or, you know,
Speaker:if you're doing Internet scale things like Google or Facebook or somebody, We're talking
Speaker:millions, tens of millions. And Right. That level of scale
Speaker:requires you know, everything has to be automated. Everything has
Speaker:to Has to work that way and it has to be resilient and it has
Speaker:to, you know, have automatic fail over and stuff. You know, there's the,
Speaker:x k CD where they're, You know, they get to a certain point. They're just
Speaker:roping off entire data centers and being like, alright. We're throwing that one away and
Speaker:moving on to the next one. And for AI, I
Speaker:think a lot of the same stuff is happening. When, you know, 10 years ago
Speaker:or so when when people were just getting started on this journey And as an
Speaker:entity, as a business entity, if you're talking about 1 or 2 use
Speaker:cases, you know, you can have humans curate that stuff and hand
Speaker:craft it, hand roll it, hand deploy it, and hand manage it. But
Speaker:if you're a a big enterprise company and you you wanna have hundreds of use
Speaker:cases in production or thousands or tens of thousands, there's just no way.
Speaker:You have to automate it. No. That that that's a that's
Speaker:an excellent point. Like, one way I've heard to describe is that if you're
Speaker:baking a loaf of bread for your family and friends Or loads of
Speaker:bread. You can do it in your kitchen. Right? You don't have to do anything
Speaker:special, but if you're the Wonder Bread Corporation or
Speaker:Mhmm. And you wanna deliver at that scale, that's no longer an
Speaker:option. Mhmm. And I think that we're at that point where and
Speaker:correct me if I'm wrong, where I think AI and ML adoption or AI
Speaker:adoption is still new enough where there's enough of naivete out
Speaker:there of, oh, we don't need to scale to that degree. Like, we don't need
Speaker:the production line. I think I think that's ending. I think we're getting close to
Speaker:the the end of that era, but that's kind of been my yeah. I think
Speaker:so too. Yeah. Because they're they're more and more, ML
Speaker:tools in everybody's toolbox. Right? So you were talking about telcos
Speaker:routing network traffic using ML models. That's not
Speaker:gonna be 1 model. Right? Like, with latency and and
Speaker:everything else, you're gonna need, you know, Very small. Lots
Speaker:and lots of very small models deployed on every router, every top of rack
Speaker:switch, every, you know, whatever 5 gs cell phone tower,
Speaker:whatever you're talking about. There are a lot of cell phone towers. So you're
Speaker:not managing 1 model. You're managing a fleet of models, right, across
Speaker:different geos and all kinds of things. No. That's that's an
Speaker:excellent point. Sorry, Andy. That's okay. It does seem to scale like that,
Speaker:though. Right? It's almost it's almost tectonic. There's
Speaker:a whole new layer going down. You know? That's that's the new surface.
Speaker:I noticed on the website, I I popped over to wallaroo dot,
Speaker:aiwallar00.ai.
Speaker:And I noticed a familiar looking, blurb just
Speaker:below the top of page. And it's familiar to me because, I
Speaker:started off in business intelligence. I'm still working in BI.
Speaker:And there's a note, 90% of AI
Speaker:initiatives Failed to produce ROI. And I saw this
Speaker:in, you know, it's very similar number, 85% in,
Speaker:in BI back in the day. It's probably still true. So where
Speaker:does that number come from? Well, I think it reflects a lot of
Speaker:things. You know? Some of them we were just talking about and
Speaker:and where MLOps is coming from is is, a lot of the failure
Speaker:modes were teams not really working with each other. Right?
Speaker:Somebody decided we should be doing AI, so they hired the data scientist.
Speaker:And the data scientist works in the corner for a while and,
Speaker:You know, 1, they don't have access to all the data. They don't know what
Speaker:the data is, where to find it, how to access it, how to clean it,
Speaker:what it means to the business. There there are a whole set of challenges there.
Speaker:And then, you know, they may train some models and and get something, you
Speaker:know, to a point where they think that it's gonna solve a problem. But Then
Speaker:you've got to work with an IT organization to stand up infrastructure. You've got to
Speaker:work with somebody to package the model and build, you know, an API around
Speaker:it or a UI of some sort And figure out how to deploy
Speaker:it, train people on how to use it, and and actually, like,
Speaker:somehow integrate it into your business process So that it's
Speaker:it's driving business outcomes. And all of those are really tough
Speaker:challenges. And all of them require breaking down those
Speaker:silos and getting a bunch of different People within an organization to
Speaker:talk to each other and communicate and to work together to solve something.
Speaker:I don't think ML or AI is is a magic wand that you just
Speaker:wave and magically provide value to a business. You've got to really
Speaker:think about What is your business doing? And, you
Speaker:know, machine learning at at heart, it it's
Speaker:really just like a a more efficient way of
Speaker:Making decisions, you know, faster and more accurately,
Speaker:and with less human input. And so you've got to look for places where your
Speaker:business can either save a lot of money or make a lot of money by
Speaker:being able to answer a a simple question repeatedly very,
Speaker:very efficient. And that sounds easy, but in practice,
Speaker:defining a business problem is often one of the hardest parts. So now I'm
Speaker:seeing even more parallels. Uh-huh. Yeah.
Speaker:You know, that was the problem we were trying to solve, with
Speaker:business intelligence as well. So didn't mean to cut you off. Sorry about
Speaker:that. No worries. So I yeah. I think I agree with you. It it there
Speaker:are tons of parallels there. I think there are a lot of similar lessons to
Speaker:be learned, and I think we are applying them in this In this space in
Speaker:ways that we've applied them to other spaces in the past. I also
Speaker:think there are technical challenges. You know, part of it is the field is moving
Speaker:so fast. So there's just this constant stream
Speaker:of of new frameworks, new models, new techniques, and you
Speaker:have to kinda stay on top of that. You have to be careful with your
Speaker:tool selection, to make sure you're not, you know,
Speaker:going whole hog into some tool. That sounds
Speaker:great today, but it's just not flexible, and it's not gonna be able to support,
Speaker:like, all these new things that are coming out. Yeah.
Speaker:Or that company could have internal internal political
Speaker:strife, which was crazy talk. Right? Cast Absolutely. Right. Cast
Speaker:doubt on their future. Alright. That would never happen. That would never
Speaker:happen. Sorry. Yeah. You were talking about privacy, which I think is
Speaker:another key thing. Yeah. Data residency, data privacy, see data
Speaker:security. You know, all of those things matter tremendously.
Speaker:And for for a business trying to, get
Speaker:value out of AI and ML. You know, a lot of it, depends on
Speaker:having good data and, Cleaning it and curating it
Speaker:and getting it ready for things. But then it it forces the
Speaker:the organization to really kind of do an inventory. What do we have? What's useful?
Speaker:What's not useful? Well, how much do we store? How much do we not store?
Speaker:How do we comply with various regulatory
Speaker:environments? Right? GDPR is is the big one everybody, you know,
Speaker:loves to throw out there. It's it's big and it's complicated, but, you know,
Speaker:+:Speaker:that. They're covered or whatever. I think that, you know, that that
Speaker:is not only do they have a big stick, but they have a big arm
Speaker:that they can wave that stick wet. Yes. You
Speaker:know, if if a small country with, like, you know, 50 people in it, and
Speaker:that could something like GDPR, people would just walk around it. But I think
Speaker:that, a block with I've heard different numbers, but
Speaker:it's for, you know, pushing 4 to 500,000,000
Speaker:people. That's a huge that's a big enough market nobody can really ignore.
Speaker:Yeah. What's interesting is on the LinkedIn page
Speaker:for Wallaroo I love the website, by the way. I checked that out too. Thank
Speaker:you. It talks about decentralized
Speaker:networks Mhmm. And at the edge. Yes. What how would
Speaker:you define decentralized network? Yeah. This is a big new push for us that we've
Speaker:been focused on for, I mean, we've been focused on it kinda for the
Speaker:last year, but it was a lot of, development on on the back end. And
Speaker:we just released kind of our 1st edge features and product,
Speaker:in October, so it's kind of a new thing for us. But,
Speaker:As you think about ML and edge or ML and AI,
Speaker:and the the fleets of models that we talked about and all these use cases
Speaker:And, you know, telcos and and five g cell phone towers and all of those
Speaker:types of things, intersecting with data and data
Speaker:residency and privacy and security, It it really seems to
Speaker:indicate to me and and to us at Wallaroo in general that the
Speaker:future is lots and lots of models being deployed in lots of
Speaker:locations. And I think that one
Speaker:big sort of industry wide theme that I'm seeing is if the
Speaker:last 20 years, let's say, was the story of Everybody
Speaker:picking up from their colos and moving to the cloud and centralizing
Speaker:all of their IT, I think that the next 20 years are gonna be
Speaker:Not like deconstructing the cloud. I think the clouds are here to stay and they're
Speaker:gonna continue to grow, right, year over year. But there will be more
Speaker:of a push out to more edge computing environments. Cell phones
Speaker:are getting more and more powerful. Cars are getting more and more powerful. Like, there's
Speaker:more computer stuff happening, all over the place, and the compute
Speaker:available, the memory and the storage available is all through the roof compared to
Speaker:what it was 20 years ago. And, I think we're
Speaker:gonna see more push for smaller, more specific machine learning models, And
Speaker:they're gonna be pushed out to all these edge locations so that they can run
Speaker:close to where the data is. So you're not schlepping this sensitive data all over
Speaker:the Internet and other people's networks. Yeah.
Speaker:But, you know, you're taking advantage of of compute resources that you
Speaker:have local to the data and making very fast decisions,
Speaker:you know, very efficiently. So I I have to jump in
Speaker:because, you you just made me feel really good.
Speaker:About a year ago, I built a large server here
Speaker:at home, which I hadn't done in a decade. Actually, my my
Speaker:20 year old son built it. But he and he helped me with,
Speaker:with picking out the new shiny fast parts, on it because I was
Speaker:so out of practice with this such confessing.
Speaker:But, and it's really cool to see, you know, all of his All of
Speaker:his skills. He does edge. We just picked up the
Speaker:Raspberry Pis are back in stock, finally. Yep. And I just picked up,
Speaker:like, 3 for $35, You know, the 1 gig force.
Speaker:Yep. Anyway, super excited about that. One of the things I built
Speaker:at the time I built a box About a year ago, you
Speaker:couldn't do a local GPT or anything close
Speaker:to that. And I said, Eventually, we're
Speaker:gonna be able to do this. I I made that guess, and it was a
Speaker:guess. Yeah. But about 6 months later, about 6 months
Speaker:ago, All of a sudden, I started seeing these 7,000,000,000
Speaker:token machines showing up and it started clicking.
Speaker:It was like, holy smokes, you can do this. I did make one stupid mistake
Speaker:and he didn't catch me on it. I bought a 12 gig GPU
Speaker:because that's super crazy huge From 10 years
Speaker:ago. And that wasn't super crazy huge at all. No. No.
Speaker:No. But it's interesting. Now they're back now. They can run on, You know, on
Speaker:the 12 gigs. And like you said, you mentioned the CPU models. So I just
Speaker:learned a ton as I've been going through this. And, That
Speaker:it's it's very encouraging to hear that. I had not heard anybody
Speaker:say edge and running small ML models on the edge.
Speaker:That's, I mean, that's what we've been trying to do here. And I I love
Speaker:the redundant you know, the idea of a redundant array of whatevers,
Speaker:you know, MLs. It's almost like a swarm of MLs. I've heard,
Speaker:yeah. Yeah. Yeah. That's true. Right? And, you know, I think there's a lot
Speaker:of interesting stuff happening on the battlefields in Ukraine right now drones.
Speaker:And Right. That Yeah. Was also a fascinating space and
Speaker:very much, I think, heading in the direction of lots of ML running at the
Speaker:edge. It's it's funny you mentioned that. So I live in a DC area,
Speaker:and, I was at a government tech
Speaker:symposium about 2, 3 weeks ago now. And
Speaker:they were talking about that that, you know, edge is gonna be much more important
Speaker:in the future of warfare. And he said presumably
Speaker:elsewhere too. Right? He was permanent primarily a government in defense. It was definitely a
Speaker:military industrial complex, type of type of event. But he was
Speaker:explaining like, you know, in the past, you know, 20 years,
Speaker:we've not dealt with adversaries. We've
Speaker:only dealt with adversaries in in battle space conditions
Speaker:where it was, you know, we controlled the airwaves.
Speaker:Mhmm. And he, I think he used an interesting term. We
Speaker:had airspace and electromagnetic electromagnetic
Speaker:dominance. I was also like, Wow. Yeah. That was yeah. Yeah. I was, like, oh,
Speaker:that's interesting. So, like, the whole idea of these disconnected
Speaker:decentralized networks, I mean, I think
Speaker:you're I think you're spot on. It's the future for
Speaker:geopolitical reasons, but also just for, you know,
Speaker:Privacy and just kind of flexibility reasons. Yeah. The
Speaker:question I have though is, like,
Speaker:Organizations can barely manage the infrastructure they have now and barely manage
Speaker:the software they have now. What are they gonna do when the software starts Not
Speaker:thinking for itself, but, like, this becomes another workload Yeah. On
Speaker:top of that. Like, what Well, for one thing, that's why Wallaroo It
Speaker:is focused where we are, and we're trying to build this platform to help people,
Speaker:you know, with this capability of being able to deploy models and manage a fleet
Speaker:of them at the edge. Because, yeah, there aren't a lot of good
Speaker:solutions for that today. Yeah. Interesting. I I think the
Speaker:general answer to your question is probably some combination of cloud and edge.
Speaker:You know, like, it does make sense to centralize a lot of things, and it
Speaker:makes the the maintenance easier and, more efficient. And
Speaker:You can get some economies of scale and, you know, all that kind of stuff.
Speaker:But, we are gonna have to get good at managing a bunch of,
Speaker:disparate types of things in desperate locations. I think all of
Speaker:us. Interesting.
Speaker:So this is the part of the show where we'll switch over to
Speaker:The premade questions, and for your convenience,
Speaker:I will, paste that in here.
Speaker:Hopefully, paste it. And there we go.
Speaker:So You had an interesting career looking at LinkedIn. You were at
Speaker:SendGrid. You were then you were at DataRobot, and you said you made a switch
Speaker:into the the data world, which begs the question, How did you
Speaker:find your way into data? Did data find you or did you
Speaker:find your way to data? I I
Speaker:guess that is a good question. I think that, it was probably a
Speaker:little bit of both.
Speaker:Finding my way to data, I think that the beginning of the story is probably
Speaker:at SendGrid. And I joined SendGrid as a DevOps engineer.
Speaker:And to be honest, I had not really heard of SendGrid at the time. I
Speaker:knew a little bit about it, but it, you know, I didn't really understand what
Speaker:it was, too much with the scale. SendGrid, by the way, is now owned by
Speaker:Twilio. But they have an API for sending email, and
Speaker:they make it just really easy to integrate with, websites and applications
Speaker:and and software so you don't have to worry about SMTP and, you know,
Speaker:DKIM signing and all the other, like, gnarly bits of of
Speaker:email. Turns out that Sengrid had a
Speaker:ton of data. They're handling billions of emails a day,
Speaker:and, you know, there's a lot of metadata there. The the actual data of the
Speaker:email and and so on, the recipients and who to send it to and all
Speaker:that stuff. And so working in that space,
Speaker:I was dealing with tons and tons and tons of data. I mean, we
Speaker:had, we were using mostly MySQL, and we had these
Speaker:massive massive clusters. I think we had,
Speaker:like, 30 or 40, you know, schemas under management. Each
Speaker:schema was a cluster of anywhere from, Like, 6
Speaker:to 40 plus servers, Wow.
Speaker:You know, with lots of compute and everything else. So that was probably my
Speaker:1st foray into, like, really thinking about data as a first class
Speaker:citizen. And, and even to the extent of, like,
Speaker:You know, building an architecture around the data. Right? So
Speaker:that you can optimize the flow of the data, and being able to store it
Speaker:and process it and transmit it fast enough to keep up with, with the
Speaker:flow. And so, yeah, from there,
Speaker:you know, had a lot of fun, learned a lot of things about, startups
Speaker:about industry, about, DevOps and and all kinds of
Speaker:things. Management as well and leadership because that's where I first,
Speaker:started managing teams And then moved to data robot and,
Speaker:into the ML space. And then it was a whole another learning journey
Speaker:about, you know, data,
Speaker:engineering, feature engineering, transformation tools. How do you
Speaker:curate your data? And how do you really, like, know what you
Speaker:have and inventory it and, make it available
Speaker:to people within the business so that they can get value out of it.
Speaker:Interesting. Very much. So our next question
Speaker:is what's your favorite part of your current gig?
Speaker:I think it's actually, I'm gonna cheat and I'm gonna say I have 2 favorite
Speaker:things. And I I kind of always have I I
Speaker:Figured out this formula a while back, in terms of what
Speaker:motivates me. And it's one part the people that I work with
Speaker:and another part, the problems that I had yet to solve.
Speaker:So I wanna work with smart people. I I really don't like being like, feeling
Speaker:like the smartest person in the room. I much prefer to surround myself
Speaker:with people that are smarter than me and I respect and I can learn
Speaker:from. But that also, you know, I enjoy. Right?
Speaker:We spend a lot of time at work, so it helps to to enjoy the
Speaker:people that you're working with. True. So that's a big part of it. And
Speaker:then, finding tough problems, hard challenges. You know, if I
Speaker:don't have hard challenges to keep me, to keep my mind
Speaker:engaged and occupied, I start to get bored and, that's no fun. I
Speaker:prefer to to always have something new to to to, you know, be chewing
Speaker:at. So, yeah, good people, smart people,
Speaker:and hard challenges. That is that is really awesome. I feel the
Speaker:same way about about both of those things. The, for me though, I
Speaker:I, Trying to find people that are smarter than me is
Speaker:really easy. So I I enjoy that part a
Speaker:lot. Like Frank. Frank is smarter than me.
Speaker:Well, thank you. So
Speaker:we have a couple of, complete this sentence, questions.
Speaker:The first one is, when I'm not working, I enjoy
Speaker:blank. When I'm not working, I enjoy
Speaker:reading. I Enjoy movies. I go biking sometimes.
Speaker:That's part enjoyment, part exercise. You know, it's good for me, but,
Speaker:There's a lot of good, road biking in particular around Denver and a lot of
Speaker:beautiful scenery. So you can, you know, just ride for a while and find yourself
Speaker:up in the mountains or something, which is great. Yeah.
Speaker:Traveling, cooking, all these things are good.
Speaker:Our next fill in the blank is I think the coolest thing about
Speaker:technology today is blank.
Speaker:I I don't think it's necessarily something about today, but I think the coolest thing
Speaker:about technology is how it builds on itself. I remember
Speaker:Years years ago, I was studying for the CCNA exam, and
Speaker:that was such a formative moment for me to
Speaker:suddenly understand How networks worked all the way
Speaker:from the physical, you know, sending
Speaker:electricity down a copper wire, and it can be on or it can be off.
Speaker:And that's it. Right? And you can do that really, really fast. Switch from on
Speaker:to off, on to off, on to off, all the way up to,
Speaker:web 2.0 and and Ajax and, you know, Asynchronous
Speaker:JavaScript stuff happening in Google Maps. Right? And I can just drag my map
Speaker:around. It's just mind blowing. And, honestly,
Speaker:like, that That journey from the zeros and the
Speaker:ones up to Google Maps, that was, you
Speaker:know, what, 50, 60 years of,
Speaker:technology building on itself of people solving very small simple
Speaker:problems, but you add up all those small simple solutions and you get
Speaker:something incredibly complex And absolutely mind blowing.
Speaker:Excellent. Very interesting. The last, the 3rd and
Speaker:final, Complete the sentence. I look forward to the
Speaker:day when I can use technology to do blank.
Speaker:I I would love, a Personal assistant, you know, like
Speaker:Jarvis from from Marvel Comics or something or, I don't know,
Speaker:from I I'm big into sci fi and and things like that when I read.
Speaker:So, there are plenty of examples, but some kind of a smart personal
Speaker:assistant that, you know, I can chat with and it keeps track of my calendar
Speaker:and reminds me of appointments and, you know, when to call
Speaker:my dad and whatever else, stuff like that. I just think that's
Speaker:so cool. And I don't you know, with Especially with all the new LLMs
Speaker:and and GPT stuff that's happening, I don't think we're super far from that. So
Speaker:it's kind of exciting to me. No. You're right. Like, I
Speaker:you know, if you watched, you know, when I was a kid, Star Trek next
Speaker:generation was on, And the way that they were able to interact with the
Speaker:computer just through their voice. Yep. And I mean, the 1st Star
Speaker:Trek show had that too, but, like, the way the conversations I thought were more
Speaker:richer and more kinda interactive. Mhmm. Mhmm. We
Speaker:have a lot of that now. Yeah. I think some of the fundamental pieces are
Speaker:in place now. Yeah. It'll probably take a little while to put
Speaker:them all together and make it work right. But yeah. Agreed.
Speaker:So our next one is, share something different about yourself.
Speaker:But we, always remind guests that we're trying to keep our clean
Speaker:rating. Yeah. On Itunes. So
Speaker:I don't know. I think one of the more interesting things about my
Speaker:Journey is that I don't have a background, like a a degree
Speaker:in anything technical. I went to college and I got
Speaker:my undergrad Studying Greek and Latin and classics. And
Speaker:so it was mostly history, archaeology, languages, and things like
Speaker:that. And Computers have always been a hobby of mine and and I
Speaker:definitely did some computer science stuff in high school. I took 1 or 2 classes
Speaker:in college, but I didn't really make my way into that
Speaker:Professionally until a few years after college.
Speaker:And, you know, honestly, I I don't think it's hurt me at
Speaker:all. And in many ways, I think it's helped me partly
Speaker:because, you know, it it helps a lot with management and leadership, just
Speaker:to To kind of have a broad background and and understand, you know,
Speaker:different people and perspectives and and where they might be coming from.
Speaker:And I'm sure that some of the languages, you know, studying languages helped me
Speaker:picking up computer languages as well. I think there are a lot of similarities in
Speaker:In, human languages and and computer, you know, programming languages. But
Speaker:What? Yeah. But, yeah, it is somewhat unique, and I don't run
Speaker:into too many other classics majors, At, you know, tech startups.
Speaker:I could definitely see the convergence, especially now when we're talking about
Speaker:LLMs and the like. Right. You know, the the
Speaker:nearest neighbor algorithms and all of that that are that are being applied
Speaker:because my understanding is that's that's, You know, that's how that
Speaker:works as it picks the next best word Right. You know, in a in a
Speaker:sentence. And so syntax and grammar and all of the things you
Speaker:studied in-depth, That should be very helpful.
Speaker:Yeah. No. That that's awesome. There
Speaker:is that good value in,
Speaker:like a classics education. I I went to Jesuit
Speaker:High School and Jesuit College, you know. Mhmm. I was forced into studying Latin
Speaker:and things like that, like, didn't do it voluntarily. I'm not gonna
Speaker:admit that, not do that. But but like as I get older, like, it's
Speaker:definitely like, Oh, I get this. Like, you
Speaker:know, especially when dealing with a lot of lawyers, there's a lot of Latin in
Speaker:that. And so I'll hear them, like, you know, Excuse some
Speaker:words. I'm like, I think I know what that means. Yeah.
Speaker:Audible sponsors data driven.
Speaker:And you mentioned you read a lot. Do you do audiobooks and
Speaker:sci fi? Do you have any recommendations? Yeah.
Speaker:There was a really good book that I read recently. Like, this is maybe
Speaker:a year ago or something, but, best book I've read recently.
Speaker:It's, The title of the book is called Seeing Like a State,
Speaker:and it's by, James c Scott. The the longer
Speaker:subtitle is, something like how Some
Speaker:schemes to improve the human condition have failed or something like that. But,
Speaker:it talks about this concept of legibility and how a lot of
Speaker:The developments over the course of the enlightenment, the industrial revolution,
Speaker:and, in the last few 100 years in in
Speaker:Our society have been primarily
Speaker:driven by the centralization of power in states
Speaker:And the state needing to administer all of these people,
Speaker:taxes, lands, land ownership, and all these different things.
Speaker:And, you know, as part of, like, the the enlightenment, the
Speaker:scientific revolution, we all got very enamored with, like,
Speaker:rational thought and Logic and and all of this stuff. And
Speaker:we thought, we're understanding the principles of the universe. We can predict
Speaker:the motions of the planets and all these things. Well, we can solve all these
Speaker:problems about, you know, around human civilization and humans as well.
Speaker:And in a lot of cases, it failed. Right? And we didn't know as much
Speaker:as we thought we did. And one of the sort of basic,
Speaker:like, premises of the book, I guess, or arguments that it's trying to make is
Speaker:that we routinely Underestimate, the
Speaker:complexity of the natural world and how necessary it is.
Speaker:And we think we can Simplify things and strip out all these
Speaker:variables and go, you know, monocultures in our in our agriculture,
Speaker:for example, and do industrial scale agriculture. You need
Speaker:timber for building ships. Great. We'll just plant Norwegian pines in straight
Speaker:rows. This is gonna be great. It's so predictable. We know exactly what,
Speaker:You know, an acre of that will yield after 10 years.
Speaker:But it turns out you can't strip out all the variables because the whole thing
Speaker:falls apart. You need the complexity of the ecosystem to keep all those trees
Speaker:healthy. And so all that predictability you thought you had
Speaker:disappears within a couple of generations because, it can't
Speaker:sustain itself. Wow. So, anyway, it it's a very, like,
Speaker:complicated book. I'm not really doing it justice,
Speaker:but I definitely recommend it. Interesting. It's on Audible.
Speaker:Yeah, yeah, so definitely check it out.
Speaker:The show. So if you go to the date is ribbon book.com,
Speaker:you'll be routed to an Audible page. And if you choose to get a subscription,
Speaker:to Audible. You will give us
Speaker:you'll get a free book, and then we'll get like a little bit of a
Speaker:bump on the head, and pat on the back, and Probably enough to
Speaker:buy a cup of coffee. It started Which will share. Which will
Speaker:share. Yes. Yes. And the final question,
Speaker:Where can people learn more about you and Wallaroo? And they even
Speaker:made that rhyme. Yeah. Great. I
Speaker:think the best place to go is the Wallaroo website, which, as
Speaker:Andy mentioned earlier is wallaroo.ai. So wallar00.ai.
Speaker:And we've got a ton of great stuff on there. Lots of, you know,
Speaker:documentation and and white papers and, tutorials and things about the
Speaker:product and what we're doing there. And for myself,
Speaker:I'm on LinkedIn. That's probably the easiest place to find me, Chris McDermott.
Speaker:And, I think I even have that as my, like, LinkedIn
Speaker:Profile name or whatever sits in the, you know, in the URL.
Speaker:Cool. It is, actually c s m
Speaker:c s McDermott. Okay. Well, thank you. Close. I was just looking at
Speaker:it, and I was also looking at the website. It's a very nice website. Thank
Speaker:you. Great design. And, although I can't design
Speaker:great websites, when I look at one, I can tell whether it's great or
Speaker:not. Me too. Me too. Same boat. I can't do it myself, but I definitely
Speaker:appreciate it. I I can't cook, but I appreciate a good meal. There
Speaker:we go. Yeah. That's it. And with that, we'll let
Speaker:Bailey finish the show. Thanks, Frank and
Speaker:Andy. And thank you, Chris, for putting up with our broken
Speaker:calendaring system. Satya should really look into that
Speaker:now that the drama around open a I is over.
Speaker:Well, over for now at least. Maybe g p t
Speaker:five can fix it.