Skip to content
Exploring Machine Learning, AI, and Data Science

Steve Orrin on the Importance of Hardware in AI Development

On this episode of Data Driven, the focus is on hardware from AI optimized chips to edge computing.

Frank and Andy interview Steven Orrin, the CTO of Intel Federal.

Intel has developed new CPU instructions to accelerate AI workloads, and FPGAs allow for faster development in custom applications with specific needs. The speaker emphasizes the importance of data curation and wrangling before jumping into machine learning and AI, 

Links

Moments

00:01:59 Hardware and software infrastructure for AI.

00:07:18 AI benchmarks show importance of GPUs & CPUs

00:14:08 Habana is a two-chip strategy offering AI accelerator chips designed for training flows and inferencing workloads. It is available in the Amazon cloud and data centers. The Habana chips are geared for large-scale training and inference tasks, and they scale with the architecture. One chip, Goya, is for inferencing, while the other chip, Gaudí, is for training. Intel also offers CPUs with added instructions for AI workloads, as well as GPUs for specialized tasks. Custom approaches like using FPGAs and ASICs are gaining popularity, especially for edge computing where low power and performance are essential.

00:19:47 Intel’s diverse team stays ahead of AI trends by collaborating with specialists and responding to industry needs. They have a large number of software engineers focused on optimizing software for Intel architecture, contributing to open source, and providing resources to help companies run their software efficiently. Intel’s goal is to ensure that everyone’s software runs smoothly and continues to raise the bar for the industry.

00:25:24 Moore’s Law drives compute by reducing size. Cloud enables cost-effective edge use cases. Edge brings cloud capabilities to devices.

00:31:40 FPGA is programmable hardware allowing customization. It has applications in AI and neuromorphic processing. It is used in cellular and RF communications. Can be rapidly prototyped and deployed in the cloud.

00:41:09 Started in biology, became a hacker, joined Intel.

00:48:01 Coding as a viable and well-paying career.

00:55:50 Looking forward to image-to-code and augmented reality integration in daily life.

01:00:46 Tech show, similar to Halt and Catch Fire.

Key Topics:

Topics Covered:

– The role of infrastructure in AI

– Hardware optimization for training and inferencing

– Intel’s range of hardware solutions

– Importance of software infrastructure and collaboration with the open source community

– Introduction to Havana AI accelerator chips

– The concept of collapsing data into a single integer level

– Challenges and considerations in data collection and storage

– Explanation and future of FPGAs

– Moore’s Law and its impact on compute

– The rise of edge computing and its benefits

– Bringing cloud capabilities to devices

– Importance of inference and decision-making on the device

– Challenges in achieving high performance and energy efficiency in edge computing

– The role of diverse teams in staying ahead in the AI world

– Overview of Intel Labs and their research domains

– Intel’s software engineering capabilities and dedication to open source

– Intel as collaborators in the industry

– Importance of benchmarking across different AI types and stages

– The role of CPUs and GPUs in AI workloads

– Optimizing workload through software to hardware

– Importance of memory in memory-intensive activities

– Security mechanisms in FPGAs

– Programming and development advantages of FPGAs

– Resurgence of FPGAs in AI and other domains

Key Facts about the Speaker:

– Background in molecular biology bioresearch

– Transitioned to hacking and coding

– Started first company in 1995

– Mentored by Bruce Schneier

– Joined Intel in 2005

– Worked on projects related to antimalware technologies, cloud security, web security, and data science

– Transitioned to the federal team at Intel

Transcript
Speaker:

On this episode of data driven Frank and Andy interview

Speaker:

stephen Oren, the CTO of Intel Federal

Speaker:

yes. Intel, the computer chip company. Because if you want

Speaker:

to train your AI models in a reasonable amount of time, you need better

Speaker:

hardware. Well, it turns out that intel has developed new

Speaker:

CPU instructions to accelerate AI workloads

Speaker:

FPGAs allow for faster development in custom

Speaker:

applications with specific needs. Speaking of intel,

Speaker:

you have to check out an upcoming intel and Red Hat webinar

Speaker:

link in the show notes. Tell them Bailey sent you.

Speaker:

Now on with the show.

Speaker:

Hello and welcome to Data Driven, the podcast where we explore the emergent fields

Speaker:

of data science, data engineering, and of course,

Speaker:

artificial intelligence. As with me, I always have Andy

Speaker:

Leonard, my most favorite data engineer in the world.

Speaker:

And today we have a special guest, Steve Oren, who is the federal

Speaker:

CTO of intel. Yes, that's right, intel, the chip

Speaker:

company. And although they do a lot more stuff

Speaker:

now. So welcome to the show, Steve.

Speaker:

Thank you and glad to be here. Frank and Andy cool.

Speaker:

So one of the things that I think people have not realized, people

Speaker:

think that AI is a software story, right?

Speaker:

Primarily. But quickly, once you get into it,

Speaker:

everyone goes gaga for things like Chat GPT or

Speaker:

well, no one's really gone gaga for Barred just yet. We're going to give that

Speaker:

a few more time for the paint to dry on that.

Speaker:

But quickly, I think when people start

Speaker:

becoming builders of AI tools, the

Speaker:

number one restriction, aside from kind of what your data engineering pipeline looks

Speaker:

like, is how quick you can train these models. And

Speaker:

obviously, I'm pretty sure intel has a thing or two to say about

Speaker:

hardware. Absolutely. And as you've as you've

Speaker:

alluded to AI, and all the things that make up

Speaker:

AI rely heavily on the infrastructure that you're

Speaker:

training you're inferencing. But even before you get to the fun stuff, how do you

Speaker:

do the data curation? How do you suck in the data? The ingestion get the

Speaker:

large multi node data sets that these large language models are

Speaker:

trained against. There's a lot of hardware and infrastructure that has to make

Speaker:

that happen. And then when you get to the important phase with how do

Speaker:

you train those in a timely fashion, hardware is

Speaker:

the answer. And what we're seeing in a lot of these spaces,

Speaker:

especially we start looking at things like large language models and transformers

Speaker:

as well as looking at other approaches that are coming out,

Speaker:

is that not only does the hardware matter, but the type of hardware

Speaker:

matters. If you think about it, it's not a one size

Speaker:

fits all. It's a heterogeneous architecture to make sure you have the right

Speaker:

hardware for your workload. One great example. So

Speaker:

large language models in graph analytics requires not just

Speaker:

heavy duty hardware but the right memory architecture to keep those nodes

Speaker:

in place while you're training. And what you find is that often

Speaker:

doesn't fit well. Intel just a classic GPU only kind of mode,

Speaker:

which is what the classic AIS leveraged, just the sheer number

Speaker:

of cores that you would have in a GPU. And so what we're seeing is

Speaker:

optimizing the hardware for the kind of workload is the answer to getting

Speaker:

timely training. And especially when you start doing more. That sort of iterative. And

Speaker:

feedback training, it's not a one and done, it's an ongoing process. So you need

Speaker:

that to be quick enough and powerful enough and robust enough to handle those

Speaker:

workloads. And then the other side where hardware really starts to matter is on the

Speaker:

inferencing, you want to be able to ask the question and get a response fairly

Speaker:

quickly, if not near real time. If you're in a car and it's

Speaker:

autonomous driving, you want it real time. You want to know that's a tree and

Speaker:

not a shadow. If you're talking about online and doing some

Speaker:

fun stuff with chat GBT, you still don't want to wait 20 minutes for your

Speaker:

response. And so inferencing matters, training matters, and

Speaker:

the kind of hardware and infrastructure that support it. And that's why intel

Speaker:

and our ecosystem are looking at providing a

Speaker:

heterogeneous set of architectures. So our classic CPU, so the Xeon and

Speaker:

the server and CPU and the client core, but also FPGA

Speaker:

based logic AI accelerators like our Habana chips in

Speaker:

the cloud and our targeted edge AI

Speaker:

chips like Movidius for video processing and the like. But then

Speaker:

really, besides the hardware, it's that software infrastructure layer. How do you

Speaker:

optimize your code? Because most AI developers are not hardware

Speaker:

experts, nor do I want them necessarily to be. So a lot of it is

Speaker:

about building out those abstraction layers that optimize your code, that's

Speaker:

doing your hugging face or whatever, to take full advantage of the

Speaker:

hardware underneath you, without you having to know what hardware is underneath you so

Speaker:

that you can provision your workload where it needs to go and not have to

Speaker:

worry about the hardware infrastructure. And that's part of our overall strategy. And working

Speaker:

with the broader ecosystem, the open source community, the

Speaker:

commercial providers, and the software frameworks to give them the

Speaker:

tools to get the best performance out of their AI and their

Speaker:

data science, right? And I think you hit the nail on the head. I

Speaker:

think we're at an inflection point. Not so much in engineering,

Speaker:

right, but more in the perception, right? Because whenever you think, oh,

Speaker:

we have a large workload we got to do, let's throw some GPU at

Speaker:

it, right? And it's a little more nuanced than that. I think

Speaker:

people are finding out that you need more than just a

Speaker:

bunch of GPU. And I was on a call

Speaker:

and I want to get your thoughts on this, because he said something very similar

Speaker:

to what you said. You ever have these moments

Speaker:

when you're on a call and somebody smart says something, you're like, I don't know

Speaker:

about that, right? And it's kind of like what they did

Speaker:

in World War Z and where there was like the 10th Man Rule, where no

Speaker:

matter how ridiculous it sounds at first, you kind of want to

Speaker:

investigate it. And that's why I was glad when your

Speaker:

name popped up in the feed because I'm like, yeah, I want to talk to

Speaker:

you about this. Because he was basically saying that

Speaker:

GPU usage is

Speaker:

overrated and that where the real advantage is going to be

Speaker:

is going to be in software acceleration and on

Speaker:

CPU kind of optimization too, which sounds

Speaker:

a lot like what you said. And when I first heard that, my first thought

Speaker:

was, I don't know about that, but this guy's plugged in. He's a

Speaker:

big shot at Red Hat, right? He's plugged in, he knows a lot. And I

Speaker:

was like, I didn't want to just dismiss that. Like, if my cousin said that,

Speaker:

I'd be like, yeah, okay, but if this guy says it,

Speaker:

whether or not he's right, maybe yet to be determined,

Speaker:

but the fact that he believes it means that there's a trail there to follow.

Speaker:

So I've been kind of poking around at stuff. Tell me

Speaker:

about that. It sounds like there's some weight

Speaker:

behind that opinion. So Frankie, you hit it on

Speaker:

the head there. It's not that GPUs aren't important, it's just GPUs

Speaker:

aren't the only and best solution for all aspects of AI. And there are

Speaker:

certain vendors that want, again, for a variety of reasons, want GPU to be the

Speaker:

foundation for all of your AI activities. Like if you're a GPU based

Speaker:

hardware company. Exactly makes sense. But

Speaker:

when you actually go look at the benchmarks across multiple and here's the key thing,

Speaker:

across multiple AI types. So different

Speaker:

algorithmic models as well as the flow, so there's different stages. So the

Speaker:

inference versus training, ingestion and curation

Speaker:

versus the training, versus the feedback training, what you'll find is

Speaker:

that GPUs will rock for certain things and they are important for certain things,

Speaker:

both from that vendor as well as from a variety of other vendors. GPUs do

Speaker:

play a key role, but when you look at the breadth of AI activities

Speaker:

and the benchmarks associated, you actually find that a lot of really

Speaker:

good work just happens on standard commercial off the shelf CPU. And

Speaker:

actually most of the inferencing, I mean, we're talking in the 70% to 80%

Speaker:

of inferencing happens best on CPU

Speaker:

and areas like large language model and graph analytic based

Speaker:

approaches. The numbers really show very

Speaker:

clearly that it's not a core bound problem,

Speaker:

it's a memory bound problem. And so having efficient in

Speaker:

and out of memory, which is what you get from a CPU or an accelerator

Speaker:

with ample memory on board, is actually much more powerful

Speaker:

for training those types of data sets because the GPU you're dealing with that

Speaker:

latency across the bus. And that actually starts to matter when you're

Speaker:

talking about billions or trillions of node graph

Speaker:

analytics. So I wouldn't say that GPUs

Speaker:

are a dying breed. That is absolutely not the case. And there's going to be

Speaker:

a huge market for GPUs or GPU like

Speaker:

functionality. I want to be careful about that because you don't have to have a

Speaker:

discrete card. The reality is you can have GPU capabilities embedded in your

Speaker:

processor. We've already seen from intel and from other

Speaker:

architectures. The real interesting thing is making sure that

Speaker:

whatever your workload is can be optimized, like your friend said, optimized

Speaker:

through software to that hardware. So that if you are

Speaker:

running a large language model, that you're actually

Speaker:

running it on the right hardware, and that the hardware and your software know how

Speaker:

to work together to give you the best performance if you're working

Speaker:

on. I'm seeing a lot of really cool things right now around graph based

Speaker:

approaches in the memory intensive side of that

Speaker:

and the switching back and forth between that. Those

Speaker:

latencies can really come to bear when you're talking about cross bus

Speaker:

kind of communication. So having high amount of memory available directly to

Speaker:

the CPU to be able to do those training, keep all that data in flight

Speaker:

so you can train, is going to be one of the key

Speaker:

differentiators of how you can take those large angle models, apply them to

Speaker:

more than just writing cool essays by Shakespeare.

Speaker:

I think what we're going to see is things like chat, GPT, and that whole

Speaker:

category of transformer based approaches applied to just about everything, not

Speaker:

just chat, but prediction

Speaker:

approaches. And it's really about getting it the training sets to become

Speaker:

smart on those very vertical domains.

Speaker:

That's going to be a resource intensive process and it's not going to be throwing

Speaker:

a bunch of GPU or it's going to be a lot of cloud scaling and

Speaker:

it's going to be a lot of memory intensive activities. And like your friend

Speaker:

highlighted, the software is going to really matter, that it's taking full advantage

Speaker:

of the hardware to get you those performance report. Well, this reminds me a

Speaker:

lot of just patterns I've seen over the decades of

Speaker:

being in computing as a hobbyist and then a profession

Speaker:

is you see a lot of things come into the

Speaker:

fore as being very monolithic, and then people

Speaker:

realize, wait, that's really a team effort.

Speaker:

And I think about it as a baseball team, right? You don't want to put

Speaker:

the pitcher, the person who's skilled at pitching in center field, can they

Speaker:

perform there? Well, gosh, yeah, but you're wasting them,

Speaker:

right? They are tuned their whole body, their

Speaker:

desires, their motivations. They love being pitchers.

Speaker:

So put that person on the pitchers mound and you see this

Speaker:

happen. And it's in all sorts of places. We saw it, frank and I have

Speaker:

seen it over the years when the unicorns were the big

Speaker:

deal, the data science unicorns who could do data engineering and everything

Speaker:

that we've kind of broken out now into other fields.

Speaker:

And we're seeing it now in the hardware

Speaker:

and in the distribution of the separation of

Speaker:

concerns and the distribution of concerns, getting every component to do

Speaker:

what it's best at. And along with that, and I'll shut up after

Speaker:

this, is this whole idea that it's moving so

Speaker:

fast that the hardware that's going to perform

Speaker:

the task first sometimes isn't even identified

Speaker:

yet because some new approach popped into the equation. If

Speaker:

somebody tested something and went, this is great. Now whether I run it

Speaker:

and you just see that and it's on a scale now where it

Speaker:

used to be measured in years and moved to months, it's now weeks

Speaker:

and sometimes days. It's just amazing how fast this

Speaker:

is going. And not that long ago, people were predicting

Speaker:

an AI intel. Right.

Speaker:

I think Dolly kind of and the whole generative artwork

Speaker:

stuff, I think kind of like, wait a minute, there's something here. Then Dolly came

Speaker:

out and then OpenAI did the one two punch of here's

Speaker:

Dolly a couple of months later, here's Chachi BT. Now you're just seeing like

Speaker:

it's on fire. Like it's not just AI summer, it's an AI heat

Speaker:

wave. Yeah, exactly. It is. It's a full El

Speaker:

Nino. I like that. That's the

Speaker:

quotable, for sure.

Speaker:

I think one of the things I think people realized is,

Speaker:

and a lot of the thinking was that AI

Speaker:

winter was coming because we're hitting processor or

Speaker:

hardware kind of upper barriers. And I think we're

Speaker:

finding out, I think much to what you said is that it's not just about

Speaker:

throw this many GPUs at it. It's right. The entire story, the entire

Speaker:

bus matters. Right. So the shortstop matters using the

Speaker:

baseball analogy. Right. The outfielders. Right. You can't really win

Speaker:

a lot of baseball games if not everybody on the team is

Speaker:

playing at their best. Absolutely. And just to take that metaphor

Speaker:

all the way, the turf matters, too. The infrastructure that you're running

Speaker:

those specialists on, you're going to play better in different

Speaker:

fields. That's true. That's a good point.

Speaker:

I love that you took the metaphor to the next level. That's awesome.

Speaker:

I think you mentioned whether it was in the virtual green room or here something

Speaker:

called habanero. And I know you're not talking about just

Speaker:

cooking. Right. Spicy habana. Yes, habana. I'm

Speaker:

sorry. I had food on my mind, as is

Speaker:

often. What is habana? Because I've

Speaker:

heard whispers of it. I know we're recording this middle of

Speaker:

May. There's going to be some announcements at the Red Hat Summit. Well, they'll

Speaker:

probably already happen by the time this goes live, but what is

Speaker:

it? So Havana is an architecture, an AI

Speaker:

accelerator, and it's a specialty chips specifically

Speaker:

designed for accelerating AI. And it's actually two

Speaker:

chips. And the reason it's two chips is that you want, again, going back

Speaker:

to what we were talking about, you want the right hardware for the AI workload.

Speaker:

So you want to be able to have the right hardware to opt optimized for

Speaker:

training flows and a separate set of hardware

Speaker:

for cloud scale and hyperscale inferencing

Speaker:

workloads. And so that's actually what Habana is. It's a two

Speaker:

chip strategy. So habana gowdy which is out available.

Speaker:

V two is available. V One has been out for some time. If

Speaker:

you go to the Amazon cloud, you can get it today. It's also available in

Speaker:

data centers, and a lot of universities have them in their high performance computing

Speaker:

environments. And it's geared to doing that sort of scale,

Speaker:

large data set training that you would find

Speaker:

whether it be in a cloud kind of environment, a chat GPT level

Speaker:

of analytic, or in the case of high performance computing.

Speaker:

Whether you're doing climate modeling or flow dynamics, those kind of big

Speaker:

training model sets that you want to be able to do at scale. And

Speaker:

what's nice about it is that like your cloud scale, it scales with your architecture.

Speaker:

So it allows you to be able to scale up your training based on

Speaker:

the compute needs with an AI accelerator specifically tuned to

Speaker:

that. The other chip, the Goya chip, is an inferencing

Speaker:

chip. So it's again tuned for that inference. But the reason,

Speaker:

again, this is for high end cloud scale hyperscale or things like high

Speaker:

speed training, where you want to be able to do large amount of inference in

Speaker:

as near or close to real time as possible against really

Speaker:

complex kind of data flows that you're trying to do

Speaker:

the analysis of. And again, looking at the right

Speaker:

hardware, we wanted to make sure to not just meet what we call the sort

Speaker:

of the normal scale. So the kind of things you would interact with when

Speaker:

you're going to do fraud detection, but you also want to be able to handle

Speaker:

really large scale inferencing because you're dealing with ingestion of multi data

Speaker:

sets across multiple different domains and having to be able to do that

Speaker:

inferencing in a streaming kind of mode. And that's really where the Goya chip

Speaker:

shines, is an inferencing platform that can scale

Speaker:

with the cloud. And that's really the Habana strategy is about giving you the

Speaker:

hyperscalers and high performance computing, the equivalent of

Speaker:

an AI custom chips. And that's really where Habana

Speaker:

sits. And then when you look at sort of the majority of what most

Speaker:

people will leverage in a cloud or on prem, what we've been

Speaker:

doing there is adding new instructions to the CPU. So

Speaker:

VNNI was the first really big one in AVX 512,

Speaker:

which really accelerates the math that you're doing behind

Speaker:

inferencing and training and give you those

Speaker:

instructions. That software, whether it be Intel's OpenVINO software

Speaker:

or TensorFlow or other frameworks can take advantage of

Speaker:

that math to use hardware offload to accelerate the math that you're

Speaker:

doing in your training and your inferencing workloads for most of your normal

Speaker:

kind of AI. A lot of the AI we deal with, not the high performance

Speaker:

computing style. And so you get the balance. And again, it goes back to what

Speaker:

we talked about in the beginning, the right compute for the right AI. We've also

Speaker:

introduced data center graphics because again, there are workloads that absolutely

Speaker:

make sense for a GPU besides fun gaming. And

Speaker:

that's really where you'll see GPU shine on, those kind of specialty

Speaker:

workloads that take full advantage. And a lot of the deep learning object

Speaker:

recognition ones work well on GPUs. They actually work well on other

Speaker:

kind of platforms as well. And one of the things we're seeing in the Edge

Speaker:

is a shift towards more customized approaches, whether that be using

Speaker:

an FPGA as sort of a hardware platform that you can code

Speaker:

in your algorithms to do inline inferencing, do feedback loop

Speaker:

training. And you see this a lot of times in the image processing, video

Speaker:

processing side, also in the signals processing. So whether it's five

Speaker:

G and being able to do signal quality testing or signal acquisition

Speaker:

and being able to do RF signal analysis, FPGAs

Speaker:

actually really shine for that kind of workload. Where you want to put in your

Speaker:

custom algorithm that you're going to actually test against or

Speaker:

use as part of your conditioning. And then we get to the idea

Speaker:

of what we call an ASIC. And that's where you know your workload, you

Speaker:

know you're going to be doing this kind of inference. You can actually code that

Speaker:

into a custom chip that will do just

Speaker:

audio AI inferencing or

Speaker:

do certain aspects of video coded. And this way you get the most

Speaker:

performance in a low swap. And that's the idea here

Speaker:

is you want to be able to handle everything from the pointy end of the

Speaker:

spear, the Edge sensor and give it the ability to do AI as

Speaker:

opposed to waiting for it to send the data to the cloud and get a

Speaker:

decision. You want to be able to give it something, but it also has to

Speaker:

operate at the size, weight and power that

Speaker:

you'd expect from an Edge sensor. You obviously don't have a data center power

Speaker:

system for your car, for your drone, or for

Speaker:

your camera on the streetlight. Right. That would be a very heavy to

Speaker:

fly that drone. That's okay.

Speaker:

I'm curious how you kind of manage what

Speaker:

I'm just going to make up words here, but like an innovation chain,

Speaker:

I'm thinking about like supply chain management. And I know

Speaker:

I've got experience in electronics engineering, and I

Speaker:

know some of how much it takes to go into mind you my

Speaker:

work was decades old, but this whole idea of getting

Speaker:

ahead of the curve or at least being able to predict where the

Speaker:

curve is going and how steep and when. That

Speaker:

sounds like a huge challenge for figuring out what

Speaker:

will be needed next. So what you're talking

Speaker:

about is how does a company that's building out both the hardware and the infrastructure,

Speaker:

stay ahead of, like you said, the week to week turnaround

Speaker:

in the AI world. Part of that is having a diverse team

Speaker:

of specialists. So the Intel Labs,

Speaker:ive to ten years out, is over:Speaker:

people who full time looking at process node technology,

Speaker:

security, AI data science. They're across multiple domains

Speaker:

and within each domain we have specialists in different areas.

Speaker:

One of the really I'll give you a great example. Before Chat GPT blew up,

Speaker:

I had two different of my AI specialists, one on the

Speaker:

government side and one on the performance side. Start talking to me about this thing

Speaker:

called Transformer. Like, oh, there's this really cool thing that we're seeing here, it's called

Speaker:

a Transformer. And I'm like, okay, that's interesting, and tell me more. And they explain

Speaker:

sort of how it worked. And then fast forward, six months later,

Speaker:

Chat chips BT shows up and I'm like, I know what that is because that

Speaker:

has the word Transformer. I've seen this. And again, it's about giving

Speaker:

your people the ability to go out and look. I think one of the

Speaker:

advantages of being at intel, and it's really why I've been here so long,

Speaker:

is everyone knows intel inside.

Speaker:

But there's something to that. Our chips are inside the

Speaker:

edge. Clients are inside the financial services, healthcare,

Speaker:

manufacturing, oil and gas. They're in the government system, they're in the cloud,

Speaker:

we're in the network. Which means we see workloads both current

Speaker:

and coming from all those different domains. So in some

Speaker:

respects we're on the cutting edge because we're seeing what people do because they come

Speaker:

to us, say, hey, I've got this software, I want to optimize on your hardware.

Speaker:

What does it do? Well, it does blah, blah blah blah. I'm like, okay, let's

Speaker:

help you. And then eventually that becomes open AI.

Speaker:

That's the kind of thing because ultimately every startup, every big company

Speaker:

wants to get the most out of their software and our teams. And one of

Speaker:

the things people don't realize is intel has over 19,000 software engineers

Speaker:

and a large majority of those do you know, they really divide up into three

Speaker:

areas sort of research and pathfinding, ecosystem

Speaker:

enabling, and then software development for

Speaker:

compilers, software services, software tools. That ecosystem enabling team

Speaker:

is a very robust team, it's been around for a very long time. Whose job

Speaker:

is to make Microsoft Windows rock on intel, make Oracle

Speaker:

rock on intel, make red hat rock on intel, make open source. We have

Speaker:over:Speaker:

to open source. We're actually one of the largest committers to open source

Speaker:

community and a lot of what they do is build the optimized

Speaker:

version of those Linux kernel libraries or to

Speaker:

that AI model running on intel and give it away and open source

Speaker:

it. We've created whole model zoos optimized for the variety of intel

Speaker:

architecture because we know if you can run it best on intel, you will run

Speaker:

it, and that consumes resources. We like that. But ultimately

Speaker:

it gives us they call them bell cows, if you will.

Speaker:

We're seeing those bell cows of what's coming next because they come to us and

Speaker:

they say, hey, help us. And very few see us as competition because

Speaker:

we're not going to go build the Chat GPT. We're not going to build a

Speaker:

new operating system or a new sort of predictive maintenance

Speaker:

solution. We're going to give you the architecture for you to run it

Speaker:

best. And even our OEM, whether you buy from Dell or

Speaker:

HP or from Lenovo, we don't care. You're buying intel hardware

Speaker:

inside. And so let's help you take the best advantage of those platforms. And that's

Speaker:

really been the approach from intel, is we want everyone's software

Speaker:

to work. And even with the GPU vendors, they still run on a CPU

Speaker:

platform. And so we want to make sure that that code runs best. So that,

Speaker:

again, you're driving the overall consumption. We raise the bar for everybody. We

Speaker:

raise the bar for everybody. Nice. Yeah. I

Speaker:

think there's a lot to unpack there. Right. And I think one of the things

Speaker:

you brought out, which is something that people don't, I don't think people have

Speaker:

widely realized yet that Edge is probably going to be the next

Speaker:

frontier in just

Speaker:

computing. Right. Obviously the last ten years have all been about cloud. Right.

Speaker:

But I think we're swifting as companies kind of take a look at the bills

Speaker:

and realize that lift and shift was not a

Speaker:

financially great decision. Right. Whether or not cloud is a good

Speaker:

thing or not, I think it always goes back to those two

Speaker:

words that every consultant and every It person always says it depends.

Speaker:

Whereas previously it was last ten years was

Speaker:

oh, definitely was the two words. But I think now we're realizing it depends.

Speaker:

And I think one of the drivers for this are things like autonomous systems

Speaker:

or drones or self driving cars, right. No matter how good

Speaker:

5G is, and I can tell you I know all the dead spots

Speaker:

in the DC area, but

Speaker:

if you're driving along at 60 miles an hour, 100

Speaker:

miles, 100 km/hour for our friends overseas,

Speaker:

and like you said, is that a tree? Is that a shadow? Is that

Speaker:

a person? Is that a grandma? Right. You don't want to wait on

Speaker:

the latency to come back. You want the inference or the decision to

Speaker:

be made on device. So you're really bumping up against the

Speaker:

speed of light and you're talking nanoseconds, not

Speaker:

milliseconds. Right.

Speaker:

What do you see? Because you mentioned you want there to be

Speaker:

sensors, but obviously these things have to be relatively low power. I guess in

Speaker:

a car it doesn't matter as much, but certainly on a drone that

Speaker:

matters.

Speaker:

What sorts of challenges does intel see in that regard in terms

Speaker:

of you want the most performance, but you want the most

Speaker:

energy efficiency. That seems like two

Speaker:

opposing forces. You would think that way, but if you

Speaker:

look at Moore's Law and you look at what's really behind that, it's about

Speaker:

reducing the size. And really that means the

Speaker:

power and increasing the performance, increasing the amount of

Speaker:

transistors. And that's really been what's driving compute all along, is how do we get

Speaker:

to lower power per density. Now, where it

Speaker:

becomes interesting is in the cloud. It's a cost measure. It's about getting

Speaker:

more for your dollar in a car or in a

Speaker:

drone or even in a factory floor. It's about being able to

Speaker:

operate closer to where the decision needs to be made

Speaker:

without having to, again, to have to power it and have that immense

Speaker:

cost. Or in the case of a drone, the weight of the battery pack and

Speaker:

so forth. So lower swap actually enables those edge use

Speaker:

cases. And again, one of the things that people realize is that Edge can mean

Speaker:

different things to different people. You talk to the cloud providers and Edge is just

Speaker:

a couple of racks closer out of the cloud. On

Speaker:

Prem, you look at Azure Stack or Snowball or these kind of

Speaker:

approaches. It's really about pushing pieces of the cloud closer to the edge through like

Speaker:

the core or they called it the

Speaker:

fog back in the day. You look at the edge and

Speaker:

you take a look at a Tesla, it's like a driving data center.

Speaker:

There's compute capabilities in there. A plane is a flying data

Speaker:

center. Your drones are getting to be more

Speaker:

computing. And when you move from a

Speaker:

discrete mode to a logical mode, and I've seen these already, where you have a

Speaker:

drone who actually has one processor but multiple containers, so actually running

Speaker:

multiple functions that could be thought of as different

Speaker:

applications on different nodes, but now they've all been collapsed with either virtualization

Speaker:

or container. So you can have navigation being one, you can be

Speaker:

doing object detection and mapping with another, and then be able to do sort

Speaker:

of other kinds of sensing like temperature

Speaker:

or barometer and things like that and doing analysis in

Speaker:

real time. One of the best examples that we demonstrated

Speaker:

at our last year's Fed summit was a set of drones out

Speaker:

mapping a region. They were going about their business, but they had a policy that

Speaker:

if somebody walked into a specific area of interest, let's say in front of an

Speaker:

embassy or in front of Lloyd or too long, that one of the drones would

Speaker:

be retasked and go over and investigate and do facial

Speaker:

recognition. All the things you want to do to make sure, hey, is this person

Speaker:

up to no good? And it didn't require a reprogramming

Speaker:

of a drone. It didn't require a special drone that was just the investigator. It

Speaker:

would basically retask itself with a new. Mission in real time

Speaker:

and go investigate. And when the person left that zone, it go back to its

Speaker:

day job of mapping the environment. That's just sort of the tip of

Speaker:

that simple prototype to show that even a very

Speaker:

small autonomous system and these were like sort of my mini drones

Speaker:

here, is capable of the compute necessary to

Speaker:

do multimission kind of use cases. So the edge absolutely is

Speaker:

that new frontier. And it's again similar to the cloud. When you say cloud,

Speaker:

everyone thinks, oh, public cloud, really? Cloud is all those architectures

Speaker:

all the way down to the edge. It's the way we develop those cloud native

Speaker:

apps that can flow back and forth. So from a cloud provider, it's moving

Speaker:

more of their cloud infrastructure closer to the edge. And what the

Speaker:

edge, folks, whether it be the actual device or sensor manufacturers

Speaker:

are looking at, is bringing some of those cloud

Speaker:

capabilities to their device to operate

Speaker:

independently. And there's a reason for that is that, number one, latency, like you

Speaker:

mentioned, Frank, but also the cost of shipping all that

Speaker:

data. No one wants to ship Raw 4K video feeds to the

Speaker:

cloud just to be able to tell me, is that a tree?

Speaker:

You want to be able to send the results that I saw a tree

Speaker:

here with the longitudinal latitude, which is a small data

Speaker:

packet, and let the sensor do the AI, do the inference

Speaker:

at the edge. Right. And then you have the case

Speaker:

where you're talking about planes or vehicles, right?

Speaker:

Like the whole time it's tracking, did the wheel fall off? Did the wheel fall

Speaker:

off? Did the wheel fall off? Right, but at one point when you get to

Speaker:

your destination, the wheel either fell off or it didn't. Right.

Speaker:

So you collapse that entire thing

Speaker:

to one integer level or really not even an

Speaker:

integer. Like a bit. Right, a bit. And then if the wheel does

Speaker:

fall off, I'm sure there's plenty of other stuff you can pick up too,

Speaker:

but hopefully nobody gets hurt. But I mean,

Speaker:

ultimately you're right. The problem with data is so much

Speaker:

that there's value, but there's a certain

Speaker:

amount of we've gotten to the point where

Speaker:

just because we can, we've done it. Right. Yeah, sure. Bring up that

Speaker:

4K. If I'm a salesperson for one of those cloud

Speaker:

providers. Yeah, man, bring in all that 4K data you want,

Speaker:

we'll take it all. We'll be happy to charge you for it too. Right,

Speaker:

but I think as we get to the point where

Speaker:

there might just be too much data, I think people organizations are going to start

Speaker:

thinking like, where can we scale back on the storage? Because

Speaker:

we don't really need it unless there's some kind of regulatory reason for

Speaker:

it. Now, one thing I want to double click on,

Speaker:

because this is a fascinating conversation, we'd love to have you back

Speaker:

on the show at some point. What's the

Speaker:

deal with FPGA because you mentioned

Speaker:

that and this was a huge deal. So a couple of things that are

Speaker:

interesting is that I first heard about Transformers at

Speaker:

the Microsoft has this internal data science conference

Speaker:

MLADS, and they first talked about Transformers. I went into

Speaker:

the talk and ten minutes, my head went

Speaker:

boom, right? I didn't quite follow it. Somebody later on in the

Speaker:

day in the reception area was kind enough to explain it, how it

Speaker:

works. And one of the other things that came out of that conference was talking

Speaker:

about the importance of FPGAs and what they're going to be like in the future.

Speaker:

Now, again, I'm a data scientist. I really don't focus on

Speaker:

hardware so much until when I need to buy new

Speaker:

hardware, like a new desktop or laptop.

Speaker:

What are FPGAs? And I remember hearing a lot about them and then

Speaker:

they kind of went dark for a while and then now they're kind of coming

Speaker:

back into vogue. Can you talk to us about, one, what they are and then

Speaker:

two where you see they're going? Sure. So Ed and FPGA are a field

Speaker:

programmable gate array. They've been around for forever. I mean, computer

Speaker:

science engineers going back, electrical engineers going back to the

Speaker:

80s played with FPGA. They were very early FPGA, but

Speaker:

basically they're programmable hardware. That's really the way to think about it.

Speaker:

You think about a CPU or an Ace or any chip it's

Speaker:

laid down with its transistors, and the flow of those transit is

Speaker:

fixed. CPU can do multiple software

Speaker:

flows, but the instruction flow is the instruction

Speaker:

flow. What makes FPGAs interesting is that you

Speaker:

can create new RTL, new layouts of flows, what

Speaker:

they call netlist of those instructions going across those transistors

Speaker:

each time. You can go in and customize it after. So the

Speaker:

manufacturing builds you a clean slate of a bunch of think about a bunch of

Speaker:

rows, and then you program them to your specific need

Speaker:

at a hardware style abstraction layer. So it gives you a much

Speaker:

faster capability because you're now really writing in hardware. It's a lot more

Speaker:

complex of a coding. It's not like doing Python,

Speaker:

but what you get is a very optimized piece of

Speaker:

hardware for your specific use case. And what's nice about that

Speaker:

is one of the great examples is in signals conditioning. When

Speaker:

you're doing like 5G research or testing signal amplitudes and

Speaker:

things like that, as you put in your algorithm actually into hardware, you go out

Speaker:

and test it. It works sort of here. I need to tweak it well, instead

Speaker:

of going and spinning a new piece of hardware, you just upload new code and

Speaker:

you go right in. So it's a much faster time of development for doing

Speaker:

those custom things. What people have found when we start looking at sort of

Speaker:

AI use cases and machine learning and pattern matching

Speaker:

is that FPGA really lend themselves well

Speaker:

to be able to create different kinds of architectural approaches to how

Speaker:

you process that data flow. If you think about a GPU

Speaker:

or CPU or even an ASIC, it's a fixed data flow. It's good for the

Speaker:

things it was designed for. What FPGA allows you to do is to customize

Speaker:

your flows based on what the data is or based on what your algorithm are.

Speaker:

And so a lot of the FPGA work they were seeing in AI is people

Speaker:

coding their AI algorithms or the machine learning algorithms right into

Speaker:

hardware and then deploying it. And so it allows you to be able to deploy

Speaker:

your thing quicker and you get pretty good performance. It's not as

Speaker:

good as say, as a custom ASIC for your algorithm. And it's not as

Speaker:

scalable really as like a software abstraction on running on a

Speaker:

cloud set of CPUs. But for a lot of these training and

Speaker:

inferencing use cases, one of the areas where it shines is in the whole

Speaker:

area of neuromorphic processing. So a whole part of the AI machine learning

Speaker:

space is modeling after brain activity or how our

Speaker:

brains process. It's a whole field. FPGAs are actually

Speaker:

well designed for those kind of algorithms that X 86 and

Speaker:

other CPU style Arctic just aren't yet.

Speaker:

And that's why FPGAs really shine in those environments, because you can create

Speaker:

these linear sort of permutation flows that you find in neuromorphic

Speaker:

algorithms. You just code those into the path for the

Speaker:

FPGA. They're really good. You'll see, FPGAs are very often used

Speaker:

in cellular and RF communications that are really good at those sort of

Speaker:

channelizer and signal optimization and

Speaker:

be able to do those kind of algorithms that you do on RF and

Speaker:

Comps, again, really good for those kind of workflows. And so why we

Speaker:

see the resurgence of FPGAs, although they've never gone away, you find them

Speaker:

everywhere. Open up your big screen flat screen TV, you'll find a couple of

Speaker:

FPGA in there. Where they're shining is because it

Speaker:

allows you to do some rapid prototyping on AI. And because we're seeing

Speaker:

now FPGAs come to the cloud. So you go to Azure has an FPGA

Speaker:

cloud. You can now deploy those algorithms at cloud scale,

Speaker:

or you can deploy an FPGA into your edge sensor and be able

Speaker:

to do that real time, sort of. Let's go try this inferencing model. Oh, we're

Speaker:

going to change the inferencing model. Let's go do that one. And where this becomes

Speaker:

really interesting in those low slop environments is a modern FPGA is

Speaker:

reprogrammable in milliseconds, which means you can go from one

Speaker:

program to another by just pushing a firmware, if you will,

Speaker:

update. And now you go from a 5G communications

Speaker:

system to LTE or to a six G

Speaker:

without actually going and swapping out the hardware. That's wild.

Speaker:

That's wild. Yeah, it's exciting times. So

Speaker:

with that, the updatable part of it,

Speaker:

how do you secure that? Because I can easily see that being like particularly

Speaker:

you work in the in the federal space, right? Like security

Speaker:

is top of mind in that work. It should be top of mind everywhere,

Speaker:

but in the near term it's top of mind, at

Speaker:

least in the federal spaces. FPGA

Speaker:

sounds like awesome, but it also sounds like that just seems

Speaker:

dangerous in a lot of ways. You can reprogram it in milliseconds.

Speaker:

There's got to be some kind of security story there. Oh absolutely. And

Speaker:

Fpjs have actually in many cases led as far as the kind of security

Speaker:

mechanisms built into the hardware for that very reason.

Speaker:

At its core, at the core level, it's the same kind of approach you do

Speaker:

for verifying your firmware on your system. It's signed

Speaker:

by hardware so that basically you're verifying

Speaker:

your load and if you're going to do an update, you're going to verify a

Speaker:

signature against a hardware rooted key so that you make sure that only

Speaker:

legitimate folks can do the update and that it's only be able to be done

Speaker:

by someone who's got the permission. From a cryptographic

Speaker:

perspective, what we find in the current FPGA that are out in the market

Speaker:

is that they've built in a whole suite of security

Speaker:

capabilities. Things like Puff Provably, unclonable

Speaker:

functions, which is basically a hardware root key that is

Speaker:

really secure as that hardware route of trust, signing in

Speaker:

cryptography functions, anti tamper functions to make sure someone can't go

Speaker:

pop open the lid or put in a jumper and try to try to change

Speaker:

the code. So those kind of mechanisms have been in place for a long time

Speaker:

because FPGAs have been used in such critical places. We find them in

Speaker:

radar stations, we find them in systems and so they've been building security

Speaker:

in for a very long time. And it's part of the workflow that when you

Speaker:

build your code you're going to take advantage of these implicit, let's call them IP

Speaker:

blocks that do security for your RTL, for your code that you're putting

Speaker:

in place. The other important thing is that the way that the code works

Speaker:

is once you lay it out, once you translate your software into that

Speaker:

layout, the layout is you can't just sort of go and reverse engineer

Speaker:

back. And so it's really a very powerful

Speaker:

mechanism as opposed to say firmware. When you're it's software.

Speaker:

If you think about the BIOS update, it's software that you're loading just deeper in

Speaker:

your platform and if anyone wants to go inspect, you'll find

Speaker:

there's a lot of software in the hardware that you don't realize is actually

Speaker:

software. The same kind of security mechanism we did there. You verify it against a

Speaker:

hardware of trust, you make sure it's signed before you run it

Speaker:

and then you apply cryptography to make sure that it can't be changed or it's

Speaker:

integrity protected. You find those same capabilities built into the

Speaker:

hardware of an FPGA and the software development tools, the

Speaker:

dialogue, the cordis and so forth have the mechanisms to take advantage.

Speaker:

So again, programmers don't have to be security gurus. They basically say,

Speaker:

I'm going to push this, and it's auto going to take advantage of those features.

Speaker:

It's good because programmers historically are very bad security

Speaker:

people. I say that. It says, yeah,

Speaker:

it's its own specialty. And yeah, you can't be

Speaker:

good at everything these days. There's too much. So I'm going

Speaker:

to echo what Frank said earlier. Steve, we got to have you back.

Speaker:

I really appreciate you being here. We could talk and geek out on

Speaker:

hardware stuff forever, but we want to

Speaker:

pivot and go to our questions and if that's

Speaker:

okay, we want to start with unless Frank, unless you had anything else you wanted

Speaker:

to do before. Let me

Speaker:

rephrase. No.

Speaker:

In the virtual green room, you talked about some things that are going on and

Speaker:

kind of operationally and

Speaker:

wow, we didn't even get there. I mean, I

Speaker:

think the important thing I took from this conversation is

Speaker:

that one, GPUs, they are important, but

Speaker:

they're not the whole story. And two,

Speaker:

at the end of the day, chat

Speaker:

GPT, any of these magical looking AI

Speaker:

models, magical seeming, right. They're all mass,

Speaker:

right? Yeah. And being beneath the math are electrons

Speaker:

bouncing around inside these microscopic chips. And

Speaker:

there's all sorts of things you could do to tweak and improve that, even if

Speaker:

it's like a billionth of a second, right? A billionth of a second times

Speaker:

a billion adds up.

Speaker:

And that adds up in terms of whether you're driving a car

Speaker:

or you're flying a plane or

Speaker:

you're a company like AWS or Microsoft,

Speaker:

where, hey, if I save one compute second per

Speaker:

transaction, I do trillions of those a day. And that's real

Speaker:

money. Exactly. And that's the thing that blew my mind. But yeah,

Speaker:

let's switch because we could geek out for hours. Because this is very

Speaker:

true. Yeah. Amazing.

Speaker:

It really is. So how did you find your

Speaker:

way into not so much data, but it how did you find your way into

Speaker:

data? Did you find it or did

Speaker:

it find you or hardware specifically? So ring. It's a really good

Speaker:

question and going back to the very beginning, actually, I started

Speaker:

out in the molecular biology

Speaker:

bioresearch side of the camp, going all the way back. I was going to be

Speaker:

a research biologist and probably still be there today,

Speaker:

except for a couple of key life events early in

Speaker:

the early ninety s, I was a hacker as a kid.

Speaker:

I loved seeing how things fell apart and how to code and break code

Speaker:

and things like that. But in the late 80s, there really wasn't a

Speaker:

career other than a COBOL programmer, which

Speaker:

wasn't an exciting career at the time. So I went the bio route,

Speaker:

which was my, the love. And right after I graduated and was going to start

Speaker:

med school, I had a year off and

Speaker:

someone had some money, wanted to do a startupy thing and they knew I was

Speaker:

a hacker and say, hey, why don't you help me get this thing running? And

Speaker:

I'm thinking, well, med school is expensive. This would be a good way to help

Speaker:

pay for it. And so I started my first company in

Speaker:

95 and after three months just fell in love with everything that was

Speaker:

going on. It was the exciting time to be in the internet. Got to apply

Speaker:

some of my security hacker background in an interesting way

Speaker:

and had some really good mentors. People like Bruce Schneier,

Speaker:

the writer of Applied Cryptography sort of took Zebru Schneider.

Speaker:

Zebrus Schneider was one of my mentors and took me under his wing.

Speaker:

And like I say, I sucked his brain dry as best as I could. But

Speaker:

really it just sort of got the opportunity to get on the ground floor

Speaker:

right before Netscape went public. So really early days on

Speaker:

a startup in the email encryption space and then one thing led to another and

Speaker:

I just felt this was what I was going to do. And for the next

Speaker:

sort of several years, I did multiple security startups throughout

Speaker:the then in:Speaker:

by intel. I like to joke, I'm still trying to figure out

Speaker:

how I ended up here for 18 years. But I think what intel

Speaker:

has provided me and provides a lot of our folks is the ability to sort

Speaker:

of innovate in an environment where a, you've got a big company

Speaker:

behind you helping you do that. But one of the best

Speaker:

reasons why I think intel has been fun for me, my most

Speaker:

successful startup, we had 500 of Fortune Thousand companies using

Speaker:

our product. The first project I worked on in intel went to 40 million

Speaker:

PCs. So the impact is just

Speaker:

unbelievable. Now from the data

Speaker:

side again, at the end of the day, like you mentioned earlier, underneath the data,

Speaker:

underneath the machine learning, underneath the AI, and even before we were talking about AI

Speaker:

was machine learning and advanced pattern matching. There's electrons

Speaker:

moving around it's running on hardware. And so a lot of what my

Speaker:

job has been before I came to the federal team was looking for ways to

Speaker:

innovate or take advantage of new use cases in software, to

Speaker:

take advantage of hardware in interesting ways. And so we call that

Speaker:

pathfinding. So you think about our labs or thinking about the next generation

Speaker:

hardware five to ten years out, I ran the team, the security

Speaker:

pathfinding team that was looking at the two to five year horizon. I

Speaker:

knew this was the hardware platform that was going to be there next year. What

Speaker:

would be some interesting things I could do with it to either advance security or

Speaker:

increase security, that was my area domain. And so things like

Speaker:

antimalware technologies, cloud security, before they knew how to spell

Speaker:

cloud. We called it virtualization security first and things like that.

Speaker:

Web security, that was the fluffy stuff. That was Steve's world while

Speaker:

the hardware engineers are figuring out low level cryptography and hardware

Speaker:

roots of trust. And we sort of worked in tandem to innovate.

Speaker:

And so as things like data science started to take off, it was like,

Speaker:

this is a key area both from a security and perspective. How do I secure

Speaker:

that data? How do I secure the algorithms? How do I use that? I mean,

Speaker:

one of the really cool things is being able to use machine learning and AI

Speaker:

and apply it to the cyber problem.

Speaker:

And when you start doing things like that, you immediately run to, well, we've

Speaker:

got too much data flowing in. I mean, the classic example is streaming

Speaker:

analytics on network at network speed. Well, how do you do

Speaker:

deep packet inspection at gigabit or higher

Speaker:

speeds without losing data? That's a big problem. That's where hardware can

Speaker:

help save you, that you just can't do in software.

Speaker:

And then when I transitioned to the federal team and took over and

Speaker:

drove our federal technology practice, you really opened the door to

Speaker:

all the different use cases. And one of the things I like about the federal

Speaker:

government is that it's a macrocosm of all verticals. You want to

Speaker:

talk finance, you've got IRS and CMS, some of the largest

Speaker:

processing of financial data. You want to talk healthcare, the VA is the

Speaker:

largest provider of healthcare, the largest insurer in the world. You want to talk

Speaker:

logistics, DoD logistics is huge. So

Speaker:

you sort of look at it, every kind of use case you'll find in government.

Speaker:

So it's really a good way of looking at all the different verticals. And they

Speaker:

all have unique or interesting data problems. There's some

Speaker:

commonality. And one of the things I really like about the federal government is that

Speaker:

you get that commonality across the divisions. They all are having trouble doing data

Speaker:

ingestion. That is just fundamental. It doesn't matter if you're the federal government or

Speaker:

Citibank or startup in Silicon Valley. Data ingestion is hard

Speaker:

and doing it at scale and being able to then do something

Speaker:

once you've got the data. And I like to use the analogy

Speaker:

of an iceberg. So AI, Chat, GPU, all these are the tip of the

Speaker:

iceberg. That's the cool, sexy stuff you can do, the hard work,

Speaker:

the data curation, data wrangling is all the work that has to be done before

Speaker:

you ever get there. And that's data ingestion, it's labeling, it's curation,

Speaker:

it's data set management, it's all that stuff. And then layer in things like

Speaker:

removing bias or dealing with bias and securing and integrity, protecting your

Speaker:

data. Like all those things have to happen before you ever start having

Speaker:

the fun math that happens towards the end of that curve.

Speaker:

That's where you find that coming out. Everyone is challenged with those things, and I

Speaker:

think that's where the excitement is today. No, you definitely hear in your

Speaker:

voice, sorry, Andy. Yeah, definitely. No, it's okay. We refer to

Speaker:

that as kind of a joke that's been going on

Speaker:

for seven years now. We say, first you get the data,

Speaker:

and that's 90% of the work. We know

Speaker:

that and your iceberg analogy fits that, Frank.

Speaker:

We need a shirt that has a picture of an iceberg against us. First you

Speaker:

get the data under the I like that. I'm definitely going to do that.

Speaker:

We launched a magazine, actually, yesterday as we record this, and

Speaker:

the cartoon segment is called First You Get the Data. And it

Speaker:

kind of like cringy things that you'll hear about data, and one

Speaker:

of them was like, yeah, first we get the data. My

Speaker:

favorite was how

Speaker:

to prep and clean the data. And they were like, oh, no, our data is

Speaker:

already in the normalized database. We don't need to clean it or prep it. It's

Speaker:

already ready. Like, oh, boy.

Speaker:

You need you need a picture of someone throwing data into a washing machine.

Speaker:

That's a good shirt. We could do that. Yeah,

Speaker:

no, that's cool. And I think you bring up something that I think,

Speaker:

folks, we don't know our exact age demographic. We have a rough

Speaker:

idea, but if there's anyone, let's say, under the age of 30,

Speaker:

right in the car with the parents

Speaker:

or they're listening, it's hard to imagine the time because we're about the same age.

Speaker:

I think you're a little older.

Speaker:

If this was not seen as a good career path, like, coding was not the

Speaker:

whole learn to code movement is a modern

Speaker:

phenomenon. I started my college career to be a

Speaker:

chemical engineer because

Speaker:

I had to convince my parents that software engineering was a

Speaker:

viable career path. And my mom, God rest her

Speaker:

souls, was like, I don't want my baby to be one of those weird

Speaker:

people in the basement. Right?

Speaker:

And then my dad, God rest his soul, was like because when

Speaker:

they came to visit me, I had a Sunday print out of the New York

Speaker:

Times, which of course had the job section, which was

Speaker:

at one point like a book. Right. And look at all these

Speaker:

jobs for computer programming. This is a thing. And my

Speaker:

dad looked through it, and he saw all the starting salaries, and it was like

Speaker:

seven or eight pages of near six figure

Speaker:

salaries in the early 90s, which was a lot of money back then, right?

Speaker:

Yeah. Like, looking through, like, on Wall Street stuff. And

Speaker:

he's like, I'm sold. And it's like

Speaker:

and my mom was like, no.

Speaker:

That is literally, like, my experience as well. When I told my parents that I

Speaker:

was going to not go to the research biology route and do the MD

Speaker:

PhD, I was going to go into the security thing. They wanted to do an

Speaker:

intervention. They thought something was wrong.

Speaker:

About two years. In 96, after I'd done the start, for about

Speaker:

a year and a half, there was an article in the New York Times, Paul

Speaker:

Cotcher, had done the timing attacks against RSA, and it

Speaker:

was front page news. And when you read down the first blurb, it says, 22

Speaker:

year old bio student from Stanford cracks RSA encryption. So

Speaker:

I cut that out and faxed it to my parents because they have an email

Speaker:

yet and said, look, another bio student doing security. It can

Speaker:

happen. Right? That's funny. One of

Speaker:

the best web developers I ever worked with, his degree was in biology

Speaker:

as well. And I think there's something to be said about understanding natural

Speaker:

systems, and I think there's some pattern matching gifts

Speaker:

that go along with that. I know my friend was that way as well. And

Speaker:

Frank, when your mom said she didn't want you to be one of those

Speaker:

weirdos in the basement that flew through my head, but I

Speaker:

maintained discipline was too late.

Speaker:

And I could say the same for me as well. Too late.

Speaker:

In her defense, my mom stayed with us in a house that my

Speaker:

wife also works in technology too.

Speaker:

She had an entire suite in our basement of our

Speaker:

house, which was not

Speaker:

windows, walk out yard, everything.

Speaker:

It worked out well. Sometimes

Speaker:

your parents my mother encouraged it without realizing. She allowed me to buy

Speaker:

the haze modem and connect it to our phone. And I did get

Speaker:

disciplined when I had that $1,000 phone bill from dialing into BBS's overnight.

Speaker:

But they should have seen it coming. Yeah,

Speaker:

my mom freaked out when I wanted a modem. She's like, no, absolutely

Speaker:

not. And my dad was like, yeah, you probably should stay out of trouble.

Speaker:

It's easy to stay out of trouble. Then. I think I was lucky

Speaker:

that my parents didn't know what a modem was, so I didn't know what

Speaker:

they were getting me. Right. This

Speaker:

is awesome. But I want to jump to question too sure. And ask, what's your

Speaker:

favorite part of your current gig? Favorite part of my good

Speaker:

gig? I think honestly, I thrive on being challenged,

Speaker:

on trying to solve big hairy problems. I think that's what has always

Speaker:

excited me is present to me with something that isn't being done well today and

Speaker:

trying to figure out how to do it. And I think one of the things

Speaker:

that I love about my job is meeting with government customers who

Speaker:

have big hairy problems and looking at a variety

Speaker:

of technologies. And I think what makes my role somewhat unique at intel, so we

Speaker:

have like a CTO for memory and a CTO for various

Speaker:

architectures is my role is pan intel so I can look

Speaker:

across FPGAs server parts,

Speaker:

networking, and sort of see that collective of where do the bits can

Speaker:

come together to solve big hairy problems. And that's really, I find

Speaker:

keeps me very excited is that every day I could be talking about an

Speaker:

IoT problem today with an edge sensor, and they're

Speaker:

talking about petabytes of data being processed in the cloud tomorrow.

Speaker:

It's looking across the technology domains and again, coming

Speaker:

from a background of cybersecurity, which again looking at various different domains from a security

Speaker:

perspective, but then adding to that AI, high performance computing,

Speaker:

it's a technology playground, right? And the federal

Speaker:

government, when I first joined Microsoft,

Speaker:

I was in the public sector, part of doing basically

Speaker:

technology developer evangelism for the federal government. And a lot

Speaker:

of my commercial sector colleagues were like, wow, it must be really boring

Speaker:

there. I might be like, you know,

Speaker:

we see things that you don't see

Speaker:

and what it is, is like there's interesting work going on, but the folks doing

Speaker:

interesting work for many reasons do not want

Speaker:

a lot of attention. Indeed. So you see

Speaker:

some things that like, wow, see, I hadn't really

Speaker:

thought of that type moments. Well, decades

Speaker:

ago I spent just a little bit of time in a really odd shaped

Speaker:

building up that way. Just a touch of

Speaker:

time. So I can have five it did. So

Speaker:

I can go yes and amen everything

Speaker:

you both have shared about. So now we have three. Complete

Speaker:

the sentences. When I'm not working, I enjoy blank.

Speaker:

Spending time with my kids. I have two small children and they keep me young

Speaker:

and full of fun and keep

Speaker:

me trying to stay in shape to keep up with them.

Speaker:

Very cool. Both Frank and I have

Speaker:

children as well. Frank has the younger kids. I'm

Speaker:

probably the old guy in this conversation now that I think about it.

Speaker:

But number two, complete this sentences. I think the

Speaker:

coolest thing in technology today is blank.

Speaker:

One thing that is a tough question,

Speaker:

I would have to say. So the two things that I think are really cool.

Speaker:

Number one, again, not the chat GPT, but

Speaker:

what the future will do with that capability is one

Speaker:

area. And then again, because I'm a security geek at heart, post quantum

Speaker:

crypto is going to be fun. Figuring out the next generation of algorithms

Speaker:

and how robust they'll be once quantum computing comes online.

Speaker:

I think that's an exciting area of math that is going to

Speaker:

spurn a lot of mathematic. Academia is

Speaker:

excited because it's a renewed interest in that space

Speaker:

and the algorithms are really interesting. The lattice

Speaker:

space structures are fun area of math to look at. Nice.

Speaker:

Interesting. The third and

Speaker:

final, complete the sentence. I look forward

Speaker:

to the day when I can use technology to

Speaker:

blank. So I'm going to give you two answers. I look

Speaker:

forward to the day when I can draw something on a

Speaker:

whiteboard and it turns into code. That's one thing I'm looking forward

Speaker:

to. Oh, nice. I can totally and that's not that

Speaker:

far off. It's not, I think a little bit of sort of the

Speaker:

image to text, image to code. I think

Speaker:

building box, you have to be able to read my horrible handwriting. That's going to

Speaker:

take an AI in its own right. But I would love a day. When I

Speaker:

can start draw my design like I like to do I'm a whiteboard kind of

Speaker:

guy, and then have it create a prototype. I think that's one thing

Speaker:

I'm looking forward to. And then I think

Speaker:

the other thing is I'm looking forward to the day when

Speaker:

augmented reality becomes reality, where it's not just

Speaker:

a cool toy, but where we actually see it integrated

Speaker:

into our daily lives. And I'm not talking to glasses and all that. I'm talking

Speaker:

about having the digital world and our physical world actually start to make

Speaker:

sense instead of it being a throwaway toy and I think we're seeing

Speaker:

pockets of it, but I think that the future is going to hold a lot

Speaker:

more of that immersive experience that we only see in movies today. I think

Speaker:

those are the two things from a technology perspective, I'm looking forward to.

Speaker:

Although I have to say, if I can get that, the code from the whiteboard

Speaker:

is going to make me a lot more efficient. No, that's true. And

Speaker:

it's funny because things that once seemed impossible

Speaker:

are now possible and even mundane. So I remember

Speaker:

when I was a kid, there was a story, there was like a story we

Speaker:

read about a kid who wrote a built a homework machine, right? And this was

Speaker:

like first or second grade and a bunch of us kids were like, yeah, how

Speaker:

do we do this? We got to make one of those. Now you look at

Speaker:

Chat GPT, obviously we abandoned the effort

Speaker:

because it just wasn't possible at the time. But you look at how kids

Speaker:

are using Chat GPU today, that machine exists

Speaker:

not in the way or the shape or form we could have imagined, but

Speaker:

it's definitely here. So to have that whiteboard to code

Speaker:

thing, it's totally

Speaker:

within sight. Whether it'll be within reach, only time will

Speaker:

tell. Probably a few weeks. If there are VCs out there listening, this is an

Speaker:

idea to invest in, for sure. I would love to see

Speaker:

especially for you, Steve. I'd love to see whiteboard

Speaker:

two FPGA code. That'd be even

Speaker:

better. We're just combining ideas. There you go.

Speaker:

I know that would make some of my engineers happy. There you go. Really

Speaker:

cool stuff. So we ask all of our guests to

Speaker:

share something different about yourself. But we caution

Speaker:

everyone to be fair that remember, we're trying to keep

Speaker:

our clean rating at itunes, so please keep that in

Speaker:

mind. So something different about me.

Speaker:

Well, I guess one thing we've already talked about that I have a bio

Speaker:

background, but the other thing I like to do is I play

Speaker:

tournament poker. I am an avid

Speaker:

poker player when not in COVID Lockdowns and things like

Speaker:d in the World Series back in:Speaker:

Really? That's something I like to do as a

Speaker:

past. It's a different use of my skills, of sort of social

Speaker:

engineering, if you will. And I like the tournament play

Speaker:

because it's sort of a long game. Right? Well, I have a

Speaker:

stack of money and I'd love to learn more about

Speaker:

is that the joke? All you need is you're always

Speaker:

welcome to my table. I'm lying about the

Speaker:

money. My wife is

Speaker:

actually a pretty good poker player, and when she was pregnant with our second,

Speaker:

she's short and she would carry a stool with her because she would have

Speaker:

to set up and her feet didn't reach the floor. And I think I gave

Speaker:

her like $100 in seed money and said, go knock yourself out.

Speaker:

And she came back like she was spending money. I think she turned that into

Speaker:

something like two grand before she had to quit and go have

Speaker:

Emma. I

Speaker:

would love to see you, because I don't think she's

Speaker:

your level by any stretch, but she did okay. We should have

Speaker:

a data driven poker tournament. We should. There we

Speaker:

go. That's an idea, Frank. The other time we had an

Speaker:

idea of somebody on the live stream said we should do like an ATV

Speaker:

race or something because we always go off track. That's kind of the joke.

Speaker:

Very true. But no, that's cool. Audible is a sponsor

Speaker:

of data driven can you recommend a good book? Ideally

Speaker:

audiobook if you do, audiobooks if not. Sure. Absolutely. Actually, I just

Speaker:

finished one that I think would be perfect sort of summation of this. So

Speaker:

Chips is an excellent book.

Speaker:

You think it's talking about today, but it gives you the history of how we

Speaker:

got here. And even one of the things I thought was really interesting is

Speaker:

some of the decisions that were made early on from the

Speaker:

policy, the government policies that we've seen and how it

Speaker:

affects where we are today. Fascinating reading. So, yes, absolutely.

Speaker:

Chips wars, it's available on Audible because I literally just finished reading

Speaker:

listening to it on Audible. So that would definitely be a book I would

Speaker:

recommend. Cool. I watched a show called Halt and Catch

Speaker:

Fire a few years ago when it was at, and it was similar. It was

Speaker:

in that vein of when things were developing and trying basically

Speaker:

the laptop development story. And of course it was

Speaker:

fiction, but I know enough about it to

Speaker:

know there were some true parallels in there. So this

Speaker:

would be very appealing to me. I'm going to get it. I hadn't heard of

Speaker:

it. Thank you for recommending and our listeners can go to

Speaker:

thedatadedrivenbook.com I didn't test it today, Frank.

Speaker:

Some days it's moody, but if you go there, it should

Speaker:

redirect you to Audible. And if you decide you get a free book on us.

Speaker:

And if you decide later to sign up, then it buys

Speaker:

Frank a cup of coffee. So when

Speaker:

you do that, we get a little bit out of it. It's a great way

Speaker:

to support the show and we really appreciate it.

Speaker:

Awesome. And where can people find out more about you and

Speaker:

what the federal team at intel is doing. So find out more about

Speaker:

me, go to my LinkedIn page. That's S-O-R-R-I-N on

Speaker:

LinkedIn. And then to find out more of what intel is doing in public sector,

Speaker:

just go to Intel.com public sector and it will redirect you to our

Speaker:

Government Solutions page. It covers everything from AI

Speaker:

data science to Cybersecurity to Edge, with lots of white

Speaker:

papers. Use cases podcasts with folks like myself and

Speaker:

others that are recording content on how intel is helping our

Speaker:

ecosystem. So definitely come check us out. Awesome.

Speaker:

And with that, I'll let Bailey finish the show. Now that was some

Speaker:

show. Is it me or are the shows getting better? It could be my

Speaker:

bias that leads me to say that, but I figured I would ask to get

Speaker:

more input. After all, what's an AI without good

Speaker:

input and a feedback loop? Speaking of feedback, have you

Speaker:

checked out Data Driven magazine yet? We are looking for writers

About the author, Frank

Frank La Vigne is a software engineer and UX geek who saw the light about Data Science at an internal Microsoft Data Science Summit in 2016. Now, he wants to share his passion for the Data Arts with the world.

He blogs regularly at FranksWorld.com and has a YouTube channel called Frank's World TV. (www.FranksWorld.TV). Frank has extensive experience in web and application development. He is also an expert in mobile and tablet engineering. You can find him on Twitter at @tableteer.