Skip to content

The Real Risks of LLMs – Guardrails, Judgment, and the Human Element in Cybersecurity

In this episode of Data Driven, hosts Frank La Vigne, Candace Gillhoolley, and BAILeY sit down with Mike Armistead, CEO of Pulse Security AI—a cybersecurity veteran who’s been fortifying digital defenses since before AI made headlines and hackers had professional profiles. Together, they dig into the dual-edged sword that is artificial intelligence in cybersecurity, exploring how AI serves as both a powerful tool against cyber threats and a potential weapon in the wrong hands.

Mike Armistead shares stories from the front lines, including his experience during the “code red” era at Google when ChatGPT shook up the tech world, and offers real-world advice on why LLMs (large language models) aren’t the magic fix for every problem—and why they desperately need guardrails. You’ll hear why your next big data breach could be hiding in a cleverly crafted AI prompt, why humans still matter when it comes to judgment calls, and why good old-fashioned security hygiene is as critical as ever.

Whether you’re a developer, data scientist, or just password-paranoid, this episode will make you rethink how you approach security in the age of AI. Tune in for expert insights, hard-earned lessons, and a few laughs as the Data Driven crew uncovers where technology, risk, and “common sense” collide.

Time Stamps

00:00 AI-Assisted Cybersecurity for SOCs

04:26 “AI Rush and LLM Insights”

09:12 AI-Powered Cybersecurity Strategy Insights

10:01 “Cybersecurity, ChatGPT, and Impressions”

13:17 AI Tools: Power and Risks

18:06 “Teaching Critical Thinking in AI Era”

20:59 “Guardrails and Next-Gen AI Systems”

24:22 Human Judgment vs AI Limitations

27:37 “Pressure Testing for Accuracy”

30:09 Future Tech Advancements and Challenges

34:58 “Risk Awareness Beyond Compliance”

37:38 “Cybersecurity Risks and AI Defense”

41:54 Cybersecurity Risks and Preparedness

43:04 “Situational Security in Practice”

46:05 “Cybersecurity’s Evolving Threat Landscape”

51:52 “Builders vs. Destroyers Mindset”

55:05 Modern Password Practices

56:39 “Pulse Security AI & Community”

Transcript
Speaker:

Welcome to Data Driven, the podcast that explores the collision of

Speaker:

data, AI and occasionally common sense.

Speaker:

Today's guest is Mike Armistead, CEO of Pulse Security

Speaker:

AI, a man who's been defending digital fortresses since before

Speaker:

AI was cool and hackers had LinkedIn profiles. We talk

Speaker:

about AI as both weapon and watchdog, why LLMs need

Speaker:

guardrails and possibly a muzzle, and how your next data breach

Speaker:

might come gift wrapped in a prompt. Grab your headphones and your

Speaker:

password manager and let's get Data Driven.

Speaker:

Hello and welcome to Data Driven Podcast. We explore the

Speaker:

emergent field of artificial intelligence, data engineering, and data

Speaker:

science. And you'll notice that Andy looks a

Speaker:

bit different today. If you're viewing the screen, if you're viewing this, if you're listening

Speaker:

to this, he'll sound a bit different. That's because Andy is actually presenting

Speaker:

a pre con today at SQL Past in Seattle. And,

Speaker:

and I am in my car because many complicated

Speaker:

reasons, but I'm not

Speaker:

driving. And I have with me my co host on Impact,

Speaker:

Quantum, which I believe that we'll all be in good

Speaker:

hands here. How's it going, Candace? It's great. It's great. I'm

Speaker:

actually really excited because with all honesty, although we focus so much

Speaker:

on Quantum, the truth is AI and Quantum

Speaker:

are now being like, spoken as if they're already

Speaker:

one word. So being able to speak today to Mike,

Speaker:

who I understand is the CEO of Pulse Security

Speaker:

AI, makes. I'm very excited about the conversation, which. Is

Speaker:

another field that is intricately tied to Quantum and AI.

Speaker:

Right. This is like, this is the center of the Venn diagram. Right?

Speaker:

So. So welcome to the show, Mike. Yeah, thank you. Thanks for having me.

Speaker:

Hey, no problem, no problem. So just a quick question.

Speaker:

What exactly does your company do?

Speaker:

You know, so we are in. I'm one of these stealth companies

Speaker:

still, so I will. But let me, let

Speaker:

me generally describe to you the problem that we're

Speaker:

addressing. And it's, it's a little bit. It's

Speaker:

interesting because I think it, it definitely falls upon

Speaker:

earlier waves of even what was going on in AI. My previous company

Speaker:

was, was called response Software. We actually used

Speaker:AI back in:Speaker:

was a little bit different though. You know, the field of AI is very broad.

Speaker:

We were probably more on the expert system end of that

Speaker:

spectrum than what. Where the LLMs are today.

Speaker:

And. But our journey was fantastic.

Speaker:

We were applying AI to do something that, I'll say it in

Speaker:

today's terms, everyone can understand, which is we were an

Speaker:

assistant for a tier one soc Analyst,

Speaker:

which if you know, in enterprises, security operations

Speaker:

centers or a soc have really

Speaker:

struggled to get skilled and even

Speaker:

just people to be able to interpret what these signals

Speaker:

are coming at them and what's a real threat and what's not a real

Speaker:

threat and what's going on there. So they have, and they, most

Speaker:

of them have to be 7 by 24. So it felt like

Speaker:

a really great application where AI, because the AI can do a lot

Speaker:

of that assistance and then give it to person, to a person to make the

Speaker:

final judgment. And we learned a lot along the way. In fact,

Speaker:

we ended up getting acquired by a company called

Speaker:

Mandiant, which is already a public company in the

Speaker:

security space. They're most known for doing

Speaker:

incident responses. So that's when you know, someone gets hacked and

Speaker:

they have to parachute in to try to get them back on their feet again,

Speaker:

which is a very, you know, manual human kind of way of going

Speaker:

through it. But they also had products and

Speaker:

our team kind of got, got involved with that. That company

Speaker:

ended up getting bought by Google a few years later.

Speaker:

And so myself and our team was at Google for

Speaker:

a couple of years. And we were at Google during an interesting time because that

Speaker:

was code red happened at Google when we were there, which is

Speaker:

chatgpt came out and Google had already had.

Speaker:

Yeah, exactly. Google already had all this

Speaker:

investment in AI that they weren't really telling anyone about.

Speaker:

And ChatGPT beat him to the punch. And

Speaker:

suddenly by edict of the CEO,

Speaker:

every product had to have AI features in it.

Speaker:

And our team was already in charge of the

Speaker:

large language model for security. And so we got to see

Speaker:

from all the teams that were doing product kind of what worked and

Speaker:

what didn't. And there's a lot because,

Speaker:

you know, I mean, it's like you guys, you see me, I've been in the

Speaker:

industry for a long time, been through many waves, was an

Speaker:

executive at a, at a

Speaker:

Internet 1.0 company in the days of

Speaker:

Web 1.0 and you know, ran

Speaker:

ops and ad tech and all this stuff from that. So I understand these different

Speaker:

waves, but LLMs aren't the answer to everything.

Speaker:

And we got to see a lot of that. Makes

Speaker:

you laugh. That's good, Frank. It should make you laugh. It's truer

Speaker:

words that's ever been said. Right. So I remember when I first made the switch

Speaker:

from Windows Phone development into AI or data science or machine

Speaker:

learning, which was called then, it was a very different world.

Speaker:

This is all pre LLM, right. I think

Speaker:

there's going to be like an ad and a BC moment

Speaker:

for AI people. It's probably going to be, you

Speaker:

know, invention of ChatGPT or the release of Chat CPT.

Speaker:

Right. And you know, where now everything's about LLMs. LLM

Speaker:

that there's plenty other types of AI out there. Right. Whether it's good old fashioned

Speaker:

math and stats, statistical analysis,

Speaker:

it's actually easier to do than it is say and,

Speaker:

or it's just, you know, old fashioned, you know, just machine learning. Right.

Speaker:

They're not related to, you know,

Speaker:

LLMs. Right. LLMs. I think it's kind of taking all the oxygen out of the

Speaker:

room for good or for bad. But I remember like I was

Speaker:

just, I was sitting at the Microsoft Research, a Microsoft Research

Speaker:

conference because I worked at Microsoft at the time and now I'm at Red Hat

Speaker:

and I'm only. They're not sponsoring this and they're not, you know,

Speaker:

approving in this or this isn't completely independent. Just want to say that

Speaker:

but my hair is a mess because haircut was one of the things I was

Speaker:

supposed to do when my hot water tank decided to blow up flood my basin.

Speaker:

Some entire weekend I've been putting stuff in dumpsters.

Speaker:

But, but like you're right, like LLMs, you know, they're great tools,

Speaker:

they're amazing. They're not going to solve everything. Right.

Speaker:

And props to Google though. The, the paper that basically made

Speaker:

LLMs, the technology on it was theirs. Attention is all you need. Was that

Speaker:

2019ish. It was, it was a

Speaker:

while ago. Yeah. You know, I mean, and a while

Speaker:

ago maybe in today's terms. In today's terms, really it's like

Speaker:

pre. Pandemic or post pandemic, honestly. Right, right. That's how people think

Speaker:

about things. Right. How things, you know, and, and

Speaker:

I. Think you know, why Google was holding on to things was there

Speaker:

was a, there was a lot of unproven

Speaker:

sides to using an LLM. And,

Speaker:

and I think, you know, so in some ways as we look at, look forward

Speaker:

and why there's so much thought about what's safe or what's not is

Speaker:

because they were kind of holding onto it. Well,

Speaker:

OpenAI didn't really feel they had that constraint and whether

Speaker:

that's a good thing or a bad thing, we're going to find out. But they're

Speaker:

both very different companies. Right. Google is a consumer enterprise company

Speaker:

and OpenAI was just a research group of at the time, maybe

Speaker:

what, 80 people, 100 people. Yeah. Right. Google was a

Speaker:

worldwide phenomena. Like so if you, if you're that big, you really have to think

Speaker:

very carefully before you release something like

Speaker:

that. Whereas if you're just a research. Yeah, yeah, for sure.

Speaker:

Anyway, to actually continue the story because that is the right

Speaker:

sidecar. No, no, no, that's fine. Because I think it's an

Speaker:

important kind of description of what's going on. We

Speaker:

then eventually we formed this company called Pulse Security

Speaker:

AI because we actually do believe that

Speaker:

there's some really great applications

Speaker:

of using an LLM. But I think within an agentix system

Speaker:

rather than just the LLM being the database,

Speaker:

we created a company that is into a place that there isn't

Speaker:

been a lot of work, which is security

Speaker:

programs are multi dimensional. There's a lot

Speaker:

to them. They grew up kind of in a technology era

Speaker:

where you solved almost one thing at a time. So if there's a

Speaker:

threat of malware, you create something that sandbox the

Speaker:

malware and detonates it and allows you to take care of it.

Speaker:

If there's a threat to access or you're over privileging things, you

Speaker:

have to think of your identity and access management. But it all kind

Speaker:

of grew up from that. But there's a layer kind of missing which is how

Speaker:

do you connect all all of this together? And security

Speaker:

teams and in our experience

Speaker:

they put people which are great on the judgment, they put people

Speaker:

in there. And so there's a lot of manual tasks about connecting the dots between

Speaker:

things. And we think AI can help a lot in at

Speaker:

the program level how you know, what people

Speaker:

should do from a strategy standpoint, not just from a

Speaker:

detailed kind of technical detection standpoint.

Speaker:

No, absolutely. My wife actually works in cybersecurity

Speaker:

for the US government at at nist. So like some of these

Speaker:

things I'm familiar with. So when you said soc, I was like, oh, I know

Speaker:

what that is, you know, like. And mand I'm

Speaker:

familiar with them, right. And it's an interesting,

Speaker:

it's an interesting time because when I

Speaker:

First, when ChatGPT released, I was just coming back from

Speaker:

reinvent in Vegas and you know, anyone's been to Vegas, right?

Speaker:

You know, like after the third day you get to the airport early because you

Speaker:

just have to get out, you know what I mean? And

Speaker:

I was starting to play with it and I was like, wow, I'm actually really

Speaker:

impressed with it. So my wife picks up the airport. I'm. All I could talk

Speaker:

about is chat gbd. Like that's literally all I could talk about. And she was

Speaker:

so like, well, I'm like, it's trained on all this corpus of data. And, and

Speaker:

she just looked at me and said, so that means all the data that it's

Speaker:

trained on is basically one giant attack surface.

Speaker:

And I was like, oh, my God, she's right.

Speaker:

But when I would tell fellow data scientists and AI engineers that

Speaker:

they would look like me, they would look at me like I had a tinfoil

Speaker:

hat on. And I was like, you know, talking about

Speaker:

conspiracies and lizard people. You know what I mean? Like, that's how they looked at

Speaker:

me. But, you know, a few years later, right, what's

Speaker:

on the owasp? It's like second or third, right?

Speaker:

Yeah, right, for sure. I mean, there's. I often

Speaker:

talk. Because you've talked about the security program,

Speaker:

it ends up that you end up talking about

Speaker:

strategy, and strategy has to include what your adversaries are doing plus

Speaker:

what you have internally. And so I end

Speaker:

up talking a bit about even

Speaker:

use of AI by the adversary and, or

Speaker:

leveraging the AI by the adversary. And so new

Speaker:

kinds of attacks based on a

Speaker:

prompt injection. I mean, that's a, that is a new thing where you

Speaker:

could, you know, just through the prompt, ask it

Speaker:

to divulge information. It shouldn't be divulging. But, but also you

Speaker:

bring up a great point, Frank, which is just

Speaker:

the LLMs, when they're getting trained, are using

Speaker:

data and you have to be very sensitive

Speaker:

to what data that is there. That's

Speaker:

why I think a lot of enterprises are scrambling to make sure that their policies

Speaker:

are set, that they can make all their employees aware of. Don't put sensitive

Speaker:

information, even though it provide great context to your

Speaker:

prompt, that it's going to be used and

Speaker:

it's going to be, it's going to be sucked in there, and before you know

Speaker:

it, it's going to be in everybody's, you know, prompt or

Speaker:

available to everybody. And it's, it's definitely a real thing. So

Speaker:

given your background in cybersecurity and talking about,

Speaker:

you know, LLMs and the LLMs adoption, what is. Do you think

Speaker:

that that is the biggest unaddressed security risk

Speaker:

is not training the LLM properly so that

Speaker:

it doesn't protect the data that it has. Like, what do you think is the

Speaker:

biggest unaddressed security risk?

Speaker:

I think a little bit related is

Speaker:

ChatGPT and Gemini and Cloud. They've kind

Speaker:

of. They're teaching everybody that

Speaker:

their system is a database of answers,

Speaker:

when, in fact, you shouldn't be thinking about it that way. You should think of

Speaker:

it as it is a tool that helps you collect the answers and

Speaker:

see the answers and do that. And

Speaker:

so the real

Speaker:

danger actually is in

Speaker:

the fact that the adversaries can use the same

Speaker:

technology to perform attacks at scale and speed

Speaker:

that we haven't really been used to.

Speaker:

And so there's that aspect to it. Then the

Speaker:

other aspect is a data

Speaker:

like hole that's there I think

Speaker:

typical in the cybersecurity world.

Speaker:

The business is really wanting to use this because it's such a

Speaker:

productivity gain and whatever it might be, either

Speaker:

your business side is either really pushing it for creative

Speaker:

work or pushing it for just understanding

Speaker:

different parts of the business. And they're ahead of the security team.

Speaker:

And that happens quite frequently. You know, in the my, not the

Speaker:

last company, but before that was all about application security. It was clear

Speaker:

software developers, you know, they were pushing the envelope about

Speaker:

making software so core to many organizations and they were thinking of building

Speaker:

stuff. They weren't thinking of somebody using it

Speaker:

to divulge, you know, corporate information

Speaker:

or to take down a corporation, you know, for

Speaker:

basically using it against them. They're creators, they don't think

Speaker:

about destroyers and the adversaries are

Speaker:

destroyers. And so you had to weave in security

Speaker:

into that culture, which remains a challenge today. I think

Speaker:

that's what's going on right now with LLMs. People are thinking, oh, I can use

Speaker:

it for all these things. They aren't thinking what it's exposing

Speaker:

Frank, back to your wife's thing. They're not thinking of the attack surface

Speaker:

you're suddenly creating by doing that. And

Speaker:

I think that's the biggest thing, Kansas, it's more that attack surface

Speaker:

expansion or just having

Speaker:

even the current attack surface just be more readily available

Speaker:

to the attackers is the thing

Speaker:

that's a real difference because ultimately it gets down to

Speaker:

even these sophisticated attacks that you're starting to hear about now

Speaker:

from the state sponsored

Speaker:

entities that are out there.

Speaker:

It's still coming down to they're exploiting age old vulnerabilities,

Speaker:

but it's just that they're getting to them in a way that's more

Speaker:

automatic. And

Speaker:

they can, as we often say in security, the bad

Speaker:

guys kind of have usually all the time in the world and they only have

Speaker:

to be right once. Yeah, right, right. Well, it's

Speaker:

interesting. You're thinking about like the jewelers, the builders, the

Speaker:

developers. That's their mindset versus the jewel heist

Speaker:

people. Right? And that's two very different mindsets.

Speaker:

And you know, I always joke that

Speaker:

our kids are going to be like the first developers ever to write secure code.

Speaker:

Right. That's my background.

Speaker:

I was a developer. But in all

Speaker:

seriousness. But one of the things that I heard is one of the things that's

Speaker:

driving companies, because you mentioned companies or businesses are encouraging

Speaker:

business users to use AI. One of the reasons I heard was

Speaker:

there was so much shadow it going on, I'm sure it's still going on

Speaker:

that if they banned it outright the stuff would just end up in

Speaker:

a public form of ChatGPT or Gemini

Speaker:

or Claude or something like that versus if they do it through the company way.

Speaker:

The companies that purvey these models, the enterprise versions,

Speaker:

they promise and pinky swear that they'll never use that

Speaker:

data for training data set in the future. So I guess that's

Speaker:

kind of better. But you're right, as I think about this,

Speaker:

we're putting AI in all of these places and we're not really

Speaker:

even sure exactly how it works. And even crazier still,

Speaker:

we're not even. Sure. We'Re not

Speaker:

even sure we know what

Speaker:

vulnerabilities are currently out there. So we're not even sure

Speaker:

now we're just pouring like all these new vulnerabilities in there.

Speaker:

We don't know what we don't know obviously. And it's just kind of like, it's

Speaker:

kind of wild like that. Yeah, I think it's also wild

Speaker:

because the AI LLMs

Speaker:

as trained, they speak so

Speaker:

authoritatively and in such, you know, proper

Speaker:

English and that you're, you're just apt to believe them.

Speaker:

You know, I, I, I, you know, one of my

Speaker:

soapbox, I guess I'll say it is that I think

Speaker:

the, one of the biggest things we can do in society today is we've

Speaker:

got to be teaching our kids at the junior

Speaker:

high, high school levels for sure and certainly in college. It should

Speaker:

be happening. But to be critical thinkers

Speaker:

because you can't, you know, if the world of social

Speaker:

media taught us anything, you know, people kind

Speaker:

of believe stuff that maybe they shouldn't believe. And,

Speaker:

and now you have an AI generating this. That sounds so

Speaker:

believable. And heck, these days, you know, you,

Speaker:

you even might see an image and think it's that person saying it.

Speaker:

It might not be at all. And yet, yet you believe that. You

Speaker:

got to, you got to kind of, you have to be, you have to

Speaker:

maybe you trust, but you got to verify. You know, it's an age old thing.

Speaker:

You just, you just can't believe things for their first blush.

Speaker:

And, and yeah, it's a whole believe.

Speaker:

H none of what you hear and only half of what you see. I think

Speaker:

now it's, you have to believe none of what you see or hear. Right.

Speaker:

Unless it happens physically in front of you. And

Speaker:

even then. Yeah, I mean look, what, what many

Speaker:

of the, the banks and other people that have to really have

Speaker:

trusted systems are doing is, you know, they're,

Speaker:

they're requiring on say a wire transfer. I know I

Speaker:

just had to do this is they, they want to call, they want me to

Speaker:

hold my, my, my license up next to my face.

Speaker:

And even then, you know, there's techniques that you use and we can get back

Speaker:

to the LLMs because you use a lot of. Well, I heard that some of

Speaker:

them will make you do this now. Yeah. Or

Speaker:

ask a question that is so off topic

Speaker:

like, and just see if, what

Speaker:

the response is, if it can't even respond, you know, ask for the favorite

Speaker:

football, pro football team or something like that, you know, and, and

Speaker:

just, you're going to be able to tell

Speaker:

using that. And if you go. So even going back. So we

Speaker:

use LLMs in our system

Speaker:

and we, we, I think the next

Speaker:

wave of things we really believe are those

Speaker:

guardrails that you have to put on it so that it won't hallucinate.

Speaker:

And you know, people think, oh, the hallucination, that's, that's an edge case.

Speaker:

It is not. You know, they, they weren't

Speaker:

really always hallucinating. I mean technically they were always hallucinating.

Speaker:

I guess you could, you could say that. I mean it's, it's a

Speaker:

probabilistic kind of way of, you know, getting the pattern and things.

Speaker:

But, but what, but what it does, it's been, the

Speaker:

models, the, the weights have been put on giving

Speaker:

a good response or a response that fulfills the

Speaker:

request and that waiting forces it

Speaker:

to make up stuff when it doesn't know.

Speaker:

And yet it sounds authoritative and things like that. And

Speaker:

so you really have to have the guardrails on it. And so I think as

Speaker:

I was saying, the next wave of systems are going to be very vertically aligned

Speaker:

like us in cybersecurity. It might be health care, it might be other things,

Speaker:

but they're going to know to make the LLM

Speaker:

to ask it basically tell me when you're like, we have, we

Speaker:

call them verification prompts, right. Or context. And so

Speaker:

it requires them to say if you're making it up, you got to tell me

Speaker:

basically, right. And then even then

Speaker:

limit what it's using as its

Speaker:

context because that'll help too. Because you can do that and

Speaker:

make it more authoritative sources rather than

Speaker:

on some Reddit board or something like that

Speaker:

where it's clearly gathering information from.

Speaker:

You have to do that. And there's people that do that. You see some of

Speaker:

the AIs being very good about

Speaker:

noting or citing its sources. I think that's something. I really

Speaker:

like it when it does that. Yeah, totally. Right, because.

Speaker:

Yeah. And they let you decide on the judgment because in my view,

Speaker:

people have to be in the middle of this for a long time.

Speaker:

Right. I'm not a believer. It's going to go sentient here

Speaker:

shortly. Again, it's my web 1.0 side to me that

Speaker:

the world was going to change. There were going to be no retailers, no bricks

Speaker:

and mortar. If you guys remember that term, bricks and mortar retailers.

Speaker:

You know, that was then they had clicking mortars. I was at barnes and

Speaker:

noble.com during, during that era. Yeah. So you know this

Speaker:

clicking water. Yeah. You know, and. But they,

Speaker:

you know, the hype was they were going to get. That was just going to

Speaker:

go by the wayside. And it was 10 years later, before that

Speaker:

really start that Amazon stopped becoming a bookstore

Speaker:

and started becoming, you know, much more than that or, or ebay got around.

Speaker:

It was much, much later. Same thing happening in AI. It's not going to.

Speaker:

These things aren't going to get there right away. So there's going to be vertical

Speaker:

use of the AI that's going to

Speaker:

provide the guardrails, provide the context that's necessary. And then people

Speaker:

start trusting those kinds of things.

Speaker:

And I think that's going to be needed for a while and then we're going

Speaker:

to see a rise of something that people then can start to trust. But

Speaker:

the LLM is not all that trustworthy right now and you need a lot

Speaker:

of stuff around it to make it accurate and

Speaker:

you know, not making up stuff. I'm

Speaker:

sorry, do you believe the future LLMs will develop

Speaker:

stronger reasoning capabilities or do you think that,

Speaker:

you know, we'll still need the human critical thinkers always,

Speaker:

you know, to close the loop. I

Speaker:

think the ultimate. We're always going to need the human

Speaker:

on judgment. So, you know, I think you can

Speaker:

close certain loops pretty accurately even

Speaker:

today with the LLM. But, but is it judgment

Speaker:

and it's, you know, the LLMs are, you know, they're just repeating patterns and,

Speaker:

and with what they have and, and things like that. So in

Speaker:

fact I just did a, you know, did a prompt recently

Speaker:

that I was asking one LLM to use another LLM

Speaker:

and it came back with kind of an odd response. So it's

Speaker:

like, so re asked it like, what version are you using? And

Speaker:

sure enough, it was using a version that was like three

Speaker:

versions ago. Because what it got trained on and

Speaker:

you just make these assumptions. It's like, oh, of course we're now

Speaker:

at ChatGPT 5. But something might not have been

Speaker:

trained on that. It might have been trained on an old version. And so there's

Speaker:

even that kind of thing happening. Sorry, Candice.

Speaker:

To fully answer your question, though, I do believe that

Speaker:

we are in for some things you might be able to close a

Speaker:

loop for. But if they involve judgment,

Speaker:

we almost ethically need to have a person involved

Speaker:

with that because you just don't know where it's going to go. And, and, and

Speaker:

you can't. And because they speak so well, people

Speaker:

are already misunderstanding that,

Speaker:

that, that, you know, they are like, like what, what they are really.

Speaker:

And they're just repeating stuff that they know. Right. They're not, they're, they're kind of

Speaker:

not making judgment calls. And there's so many things that are just

Speaker:

about judgment that I think it's just better to think of

Speaker:

them as a tool, not as this thing. I,

Speaker:

I think there's a lot to get to, to get to these, you know,

Speaker:

I know, I don't know. Sam Altman might say it's only two years

Speaker:

away. I just think that's, that's, there's no way

Speaker:

not, not for proper ethical judgment.

Speaker:

Right? I mean, yeah, it might fake it

Speaker:

really well, but it won't be ethics

Speaker:

based judgment. And so do you think we

Speaker:

could use AI tools to design better prompts.

Speaker:

That we do that all the time? Absolutely

Speaker:

you can. And in fact, I think it's

Speaker:

almost the best practice now that you are both,

Speaker:

like I had mentioned before, kind of the truth or the

Speaker:

truth directive that you give it, you can give it a lot of pros. We

Speaker:

also notice it's kind of, I don't know, like

Speaker:

what about a month ago there

Speaker:

was a very illustrative

Speaker:

way of you need to threaten these things because it'll access

Speaker:

or raise the stakes for these things because it'll access different parts of the

Speaker:

model. Back to your thing, Frank. We don't really understand how they really

Speaker:

work. And so it was just mind blowing in a way

Speaker:

that you have to say. And so we even

Speaker:

give our prompts the ability to say, hey,

Speaker:

I will lose my job if I don't get this right. So get

Speaker:

this right. But we definitely play the models

Speaker:

off on each other because it's good

Speaker:

and it's kind of asking one to be

Speaker:

the devil's advocate on the other. And that's a known

Speaker:

group think, you know, think about just people socially. Right

Speaker:

group thinks, been around forever. And the way you,

Speaker:

you go against it is you ask someone to be the

Speaker:

devil's advocate in whatever this judgment needs to be. And

Speaker:

that's a great way to test, pressure test if what you're hearing

Speaker:

is actually right or not. And so yes, we have to pressure

Speaker:

test, use the elements to pressure test each other, use our own

Speaker:

prompts to pressure test the current model. You know, there's lots of different, different

Speaker:

techniques to do this. I mean, you know, I think of your

Speaker:

world especially, you guys have long specialized in, you

Speaker:

know, data science has been one of these areas that uses a lot of these

Speaker:

techniques to make sure that it just, you know, you don't get too narrow

Speaker:

in the focus and you know, and you get to get the right answers.

Speaker:

There's a whole, there's a whole new set of things that have to be

Speaker:

done to make sure that we're, we're, we're using the tool in the way

Speaker:

we should use it.

Speaker:

I love you guys. Speechless. It's

Speaker:

interesting. Like, and what's your take on private AI? Right, like running

Speaker:

your AIs entirely on prem within servers you can control.

Speaker:

I mean, I know a lot of people, including myself, think that's

Speaker:

the cure all for a lot of these issues, but even then I'm thinking like,

Speaker:

if it sounds like a cure all or a silver bullet, it's probably not.

Speaker:

Yeah, I mean, I think it, it solves a bunch of these problems that we've

Speaker:

been talking about. You know, it clearly does, but you

Speaker:

can't air gap it totally because people, you want people to be using it.

Speaker:

And so it'll, you know, you're still going to have insider threats.

Speaker:

And so if you have an insider, you know, there's still going

Speaker:

to be ways of getting information out. And it might not be

Speaker:

a risk that you want to take as a company. I mean you still,

Speaker:

and, and so you still have certain things

Speaker:

that do it. But I do think it solves some things. The thing it doesn't

Speaker:

solve is the, you know, why we're seeing

Speaker:

such a rapid advancement in stuff is because

Speaker:

it's the LLMs are looking at everything kind of that

Speaker:

are public, that's public out there and making use of

Speaker:

those and then people are looking at them and going, oh wow, that's great. And

Speaker:

doing that, you'd have to replicate a bit of that. And yeah, you could bring

Speaker:

those in and, and but there's going to be a lot of advances that we

Speaker:

can't even predict right now. You know, like talking to, you know, Candace, you

Speaker:

on the quantum side or you know, now we're seeing that, you know,

Speaker:

Nvidia's got the chips right now, but the wafers and

Speaker:

the amount of transist, you know, transistor equivalents you can put

Speaker:

on these things are, it's going to impact things and maybe it's

Speaker:

going to be practical, you know. No, nobody thought we'd have a whole

Speaker:

computer on our phones, but you know, us going back

Speaker:

into the 80s, yeah, it was a, that's a

Speaker:

pretty powerful computer compared to what we were using at the time. You

Speaker:

know, there's a lot of those things that are going to come to play. And,

Speaker:

and so I, I do think bringing some of this stuff internal

Speaker:

and that it'll solve some things. It won't solve everything though.

Speaker:

And you know, you'll still have to, you'll still have to do a lot of

Speaker:

good security hygiene. You'll start to do a good, a lot of good data hygiene.

Speaker:

You know, I mean I'm kind of. Worrying though because like companies have not been

Speaker:

doing a really bang up job of that last 50 years.

Speaker:

Yeah, it's more noticeable, it's more noticeable now more than ever.

Speaker:

I wonder what new vulnerabilities would private

Speaker:

AI, what new vulnerable would private

Speaker:

AI solve? And what or what, what new

Speaker:

vulnerabilities would it, would it expose? Right? Like because we

Speaker:

still don't know even if it's running on your server, you still don't know how

Speaker:

it works. You know the only thing. And

Speaker:

you're right and you also, that's why I brought up the

Speaker:

insider. You know, it's an attack surface. You

Speaker:

know, maybe you closed it down a little bit from being external but you have

Speaker:

insider threat threats. You have other right. You know, the creative

Speaker:

things that are going on on the attacker side about

Speaker:

they've long kind of done, you know, the,

Speaker:

where the attacks were. They'll get inside and they'll just wait and

Speaker:

they'll wait for kind of the dust to clear so you cannot trace it back.

Speaker:

And they'll cover their tracks and, and it

Speaker:

could be sitting there in the first time someone in the business

Speaker:

connects that model that's you think is walled

Speaker:

off to something even for good

Speaker:

legitimate business reasons. It might expose

Speaker:

an avenue that someone could get in and start exfil trading. And you may not

Speaker:

even know they are, I mean these low and slow

Speaker:

Attacks that have been the bane of so many

Speaker:

enterprises where you're just siphoning it off enough so

Speaker:

that the controls don't see it. Those will happen

Speaker:

in a lot of. And they could happen to models. And there you have your

Speaker:

crown jewels, your data. That's everything slow slowly being

Speaker:

siphoned off. You know, that's, that's going to remain. And

Speaker:

you're going to have to have a multi layer security

Speaker:

system in place to kind of deal with that as well.

Speaker:

No, that's true. And it makes me wonder like,

Speaker:

you know, I guess, I guess you can

Speaker:

be. I had an actually interesting conversation with the customer a couple

Speaker:

years ago and he talked about what's called the. And I know I'm going to

Speaker:

mess up what the acronym is, but it's CIA Triad. And there's

Speaker:

nothing to do with the Central Intelligence Agency. It's

Speaker:

confidentiality, something.

Speaker:

And what is it? Probably identity. Yeah.

Speaker:

Or integrity, I think. And then access.

Speaker:

Right. And he had this whole, you know, he had a whole thing where like,

Speaker:

you know, if you lock things down so much, you basically kill

Speaker:

the access part of it. Right. You basically make it impossible to access. Right. If

Speaker:

you. It seems like security is one of those jobs that

Speaker:

will be augmented by AI for sure. Right. Because no one's going to have time

Speaker:

to read gigs and gigs of log files anymore. Right.

Speaker:

But it's also going to need. You're going to need a human in the loop.

Speaker:

Right. I don't say that because that's what my wife does. And

Speaker:

I like.

Speaker:

Yeah, I like paying. You bring up a great point.

Speaker:

And let me transition it to this because

Speaker:

I think I'm going to use a term that gets misapplied

Speaker:

a lot for enterprises and it's about risk.

Speaker:

You are not going to. And Candice, it gets to your

Speaker:

point too. The job of security

Speaker:

programs inside of enterprise is actually to mitigate

Speaker:

the risks to the business. It's not to provide 100%

Speaker:

security. That's not the goal. The goal is to mitigate the risk because

Speaker:

every business is going to have risk. And, and you need to accept a certain

Speaker:

amount of risk so that you can do business and you can reach more

Speaker:

people and you can, you can do that. And

Speaker:

circling all the way back to what Pulse Security

Speaker:

does is that we hope to bring that concept back into things, is

Speaker:

that the leaders should be thinking about risk

Speaker:

and tracking their risks and knowing where they're

Speaker:

taking risks or where they're not taking risk. I think today

Speaker:

why I said it's kind of one of these misplaced things is we kind

Speaker:

of allow regulations and things like that

Speaker:

to say to be about risk. And that's really the low

Speaker:

bar, you know, in security we always talk about, you know, if

Speaker:

you're, you talked about the OWASP earlier or you know, you think about the

Speaker:

PCI standard, you know, for retailers and

Speaker:

transaction processing, or you think about some of these other

Speaker:

standards, they're the low bar. And many people

Speaker:

think about risk as something that I have to do it we call

Speaker:

checkbox compliance. Right. I have to be compliant, but I only want to

Speaker:

do as much as I, as I can. As you need to. Because no one

Speaker:

looks forward to seeing security people, whether it's physical security, you know,

Speaker:

or it security. Like you know, the things

Speaker:

developers have told her. Right.

Speaker:

You know, and like, as a form, you know, as a former developer, like, I

Speaker:

get it, like, and I know data scientists don't think about this

Speaker:

generally speaking, right. Data engineers might,

Speaker:

but even then, like, you know, they're, I think you said

Speaker:

it earlier like it's the mindset, right? You know, you have the builder mindset,

Speaker:

maybe the plumber mindset for data engineers and

Speaker:

then you have kind of the attacker mindset, right. These are different ways of

Speaker:

thinking. It almost cries out that you need to have

Speaker:

diverse mindsets on these projects now. I mean, you always need.

Speaker:

Now it's more obvious. That's why you see, yeah,

Speaker:

that's why you see

Speaker:

good hygiene insecurity is that you have, you do

Speaker:

a threat modeling before you are going to go external. And

Speaker:

that is a very different person that usually guides that. It's, it's exactly what you

Speaker:

said, Frank. They have, they have the. I'm going to bring to you

Speaker:

this, what you might feel is a very, very big edge case.

Speaker:

But if there's a probability can happen, you have to consider

Speaker:

it and you have to, you have to think about it. And that's back

Speaker:

to something we talked about earlier. When you have the attackers using

Speaker:

AI, they can explore the

Speaker:

corners and these edge cases so much easier.

Speaker:

And if they find one. And it could be very

Speaker:

unsophisticated though, because it could still be a vulnerability

Speaker:

that has been around for 10 years and believe it or not,

Speaker:

that's still going on that the ultimate way they got in

Speaker:

was a very old unpatched

Speaker:

resource that just happened to get exposed. But it was, it was the

Speaker:

reason other term in security, the lateral movement of the bad guy, that they're

Speaker:

just, they're just moving laterally to investigate

Speaker:

different parts and they found something and they got in that way

Speaker:

and back, back to the thing. Just have to be right once and

Speaker:

poor defenders have to be right 100% of the time and they

Speaker:

won't be. So that's why taking a risk approach

Speaker:

is the way, is the way to go. Because no matter

Speaker:

what your size of the company, you got to consider your budget,

Speaker:

how many people you have, the skill set of those people.

Speaker:

And this is where I think AI can really assist the

Speaker:

defenders is that it can add some of that

Speaker:

expertise and some of that, you know, vigilance

Speaker:

that's on 24. 7 in ways that people

Speaker:

can't. But they got to bring it to people and the people can make the

Speaker:

judgment call. Because if the AI had its way, I mean,

Speaker:

Frankie kind of said this too, that the most secure thing is to

Speaker:

shut the whole thing down and not let customers access it.

Speaker:

You know, and you don't want that because that, that's your business, you know. That

Speaker:

kind of defeats the purpose. Yeah, exactly, exactly.

Speaker:

So a risk based approach is super important. And, and it is

Speaker:

about then just, you know, you judging how much

Speaker:

risk you want to take and your board wants to take and you

Speaker:

know, and the CEO wants to take and the business people want to take and

Speaker:

then, and then applying that and making sure that,

Speaker:

you know, it matches your business.

Speaker:

And so that's, you know, that, that's a lot of the game. Right.

Speaker:

That makes a lot of sense. So what is your. I'm sorry, candid. Go ahead.

Speaker:

So what would trigger the shift inside an organization

Speaker:

from reactive security to risk aligned decision

Speaker:

making? You

Speaker:

know, oftentimes, unfortunately,

Speaker:

it's you get hacked and

Speaker:

you, a lot of times also, unfortunately, you bring

Speaker:

in new leadership who understand that their charter is to

Speaker:

come in and change the culture a bit, you know,

Speaker:

from that. Now existing leadership can certainly do that, but

Speaker:

whether they're given enough chance to, I

Speaker:

don't know, I, you know, it's. All fun and games

Speaker:

until somebody gets hurt. Right. Like, and you know, and I think that

Speaker:

if you'd never had a problem before and then it suddenly

Speaker:

happens. Right. I don't think it's, there's

Speaker:

a joke, it's a bit of a gallows humor type thing where

Speaker:

like clockwork, within 24 hours of a major breach of a major company. Right.

Speaker:

What do you see? Job listings for

Speaker:

cybersecurity? Okay. There was a major, I

Speaker:

think it was one of the major hotel chains. I think you all know who

Speaker:

we're talking about. I don't want to, I don't want to name Anyone by name,

Speaker:

I don't want to get sued. But you know, literally

Speaker:

like within a week, you know, there was like two pages of job

Speaker:

listings for, you know, some flavor of

Speaker:

cybersecurity or security analysis.

Speaker:

And it's unfortunate that in many

Speaker:

organizations the security leader is kind of

Speaker:

set up to be the scapegoat when something like that happens. When

Speaker:

in fact, you know, you could be doing all the right things.

Speaker:

And I don't know, you guys probably know the term

Speaker:

too. The, a black swan event kind of happens. Which, right, which,

Speaker:

you know, we know it, everybody knows it if they travel. Because

Speaker:

how often have we caught someone who has the explosive in their shoes

Speaker:

getting on an airplane? It's never happened since the first time.

Speaker:

That, and that was a black swan event. And yet we

Speaker:

designed our whole security, a lot of our security around

Speaker:

that and that. And it should be done around the major,

Speaker:

the risks. And if you think about

Speaker:

it in that way, really if you've

Speaker:

traveled internationally, especially in places that they really have a risk,

Speaker:

they will often randomly pick a plane,

Speaker:

get everybody off, look at all the baggage. But it's a

Speaker:

random kind of thing that happens

Speaker:

rather than kind of a systemic way of going through it that becomes

Speaker:

kind of wrote and it, you know, and, and people learn how to defeat it,

Speaker:

you know, in some ways. And, and that happens in cyber security

Speaker:

all the time. You know, you gotta, you gotta really be

Speaker:

doing that. That's why doing like practice, you know,

Speaker:

it's, it's a, it's a real important thing to do what we call

Speaker:

tabletop exercises in this because you have to

Speaker:

pretend like you just got hacked. What, what do you do

Speaker:

from the lowest level analyst all the way up to

Speaker:

the board. Do they, do they know what to do? Because,

Speaker:

and there's a lot of regulations now within like 24 hours, they have to,

Speaker:

or 72 hours of detecting it, they have to, they're on the

Speaker:

hook to claim it. I forget what that law is called. Yeah,

Speaker:

that's right there. If you're a public company, you have to disclose

Speaker:

and typically you don't have any idea yet

Speaker:

what, how that's happened and yet you have to disclose it's

Speaker:

happened and it's, you know, so,

Speaker:

so yeah, there's, there's a lot of risk to the organization

Speaker:

that, that this, this presents. And so that's why having,

Speaker:

thinking about it that way, doing exercises, you

Speaker:

know, it is a new world. I'll say. I, I'm

Speaker:

a big believer in what I'm going to call situational security where

Speaker:

And I mentioned this before, you just got to know your situation and if the

Speaker:

stakes are high and you have a big security team and you better be

Speaker:

practicing these things, you better have done, you know,

Speaker:

multi layers of security. But if you're a small team and you only have a

Speaker:

couple people on it, you've got to kind of think of what your crown jewels

Speaker:

are. Go protect those first and let the other

Speaker:

stuff go because who cares who's on your guest network?

Speaker:

You know, it's like you got to let that go, maybe make sure that,

Speaker:

make sure your guest network is not tied to your internal network.

Speaker:

And I think these days you have to really look at access

Speaker:

because so much. Everyone's in the cloud with

Speaker:

a lot of their infrastructure these days and you can tell a lot

Speaker:

by, so don't over privilege people to have access to things

Speaker:

and do that. So you have to look at those kinds of things. You do

Speaker:

have to look at your, you know, it goes without saying,

Speaker:

look at, look at your resources and I'm going to use that term broadly,

Speaker:

your assets that you have because you have to know about them.

Speaker:

So having some protection on those assets is super critical as well.

Speaker:

And I'm an old appsec guy, so yes, you, and you know, Frank,

Speaker:

you mentioned you got to have your applications

Speaker:

that are actually performing much of the business these days. You have to,

Speaker:

you have to know what your vulnerabilities are and you've got to plug the big

Speaker:

holes in that. But from there really

Speaker:

can't stop the bad guys. But you can at least stop,

Speaker:

stop the amateur bad guys. Right. Well, and, and

Speaker:

they're going to look around. So you can, if your bar is

Speaker:

higher than the next guy, as we know. You know, I know

Speaker:

that's all these adages of, you know, running faster than a bear or

Speaker:

the next guy, you know, on the bear and all things. That'd be the second

Speaker:

slowest. That's right. And it is true, you know, you can, you can

Speaker:

dissuade a lot of attacks if you

Speaker:

look like it's going to be difficult because the attackers,

Speaker:

they run playbooks too, because it's easier for them, it's cheaper for

Speaker:

them. It. And they'll just run playbooks. And if, if you

Speaker:

thwart the playbook, they'll find someone who, who

Speaker:

doesn't. And if you're not state actor, it's a, it's a criminal

Speaker:

enterprise. Right. And criminals are there to make money. Right. State

Speaker:

actors have different motives and different budgets.

Speaker:

Yeah. They may go a lot more targeted and they're just going to wait and

Speaker:

be patient. You're exactly right. But

Speaker:

actually targets like. I'm sorry,

Speaker:

no, I don't mean to interrupt. I was just going to say a funny story

Speaker:

is when we were

Speaker:

pitching our application security company

Speaker:back in:Speaker:

how

Speaker:

underfunded but patient and have all the times in the

Speaker:

world the hackers are, we were kind of

Speaker:

dismissive of no state would ever

Speaker:

hack another state's assets because it start a

Speaker:

war. And at the time that was really, that was the thinking.

Speaker:

I mean, how quaint does that sound today when we all know it's like, oh,

Speaker:

that's a Russian hacker group. You know, it's like we just kind of go,

Speaker:

oh, of course it was. It's like, oh my gosh. Yeah.

Speaker:

Well, it's also, I think in terms of geopolitics become a real

Speaker:

equalizer. Right. Because a nation state like North Korea can go toe to

Speaker:

toe with the United States. Right. Whereas in a

Speaker:

conventional war really wouldn't work out well for them. You know what I mean?

Speaker:

It's an interesting, yeah, it is interesting.

Speaker:

We have really good hackers ourselves in the

Speaker:

United States, right? Oh, I'm sure we do. It's, it's, it, you know,

Speaker:

I mean I, you know, you kind of hope that but,

Speaker:

but you are right, like a North Korean thing like, like we've seen

Speaker:

they can use these deep fakes to infiltrate in ways

Speaker:

that, you know, because of the work at home thing how, how they

Speaker:

can get employees hired in some of these places

Speaker:

with the expressed intent of, you know, stealing

Speaker:

things, you know, from, from those organizations and it's,

Speaker:

yeah, it's, it's a new world. It's pretty wild. Yeah. I mean when you think

Speaker:

about it like in, and you know, and it's not, not saying that the United

Speaker:

States doesn't have good hackers. I'm sure, I'm sure we have among the best. I

Speaker:

mean, maybe the best. But it's like a

Speaker:

baseball team, right? Like, you know, obviously there are some baseball teams that are going

Speaker:

to be better than others, right. And it's going to be kind of like the

Speaker:

smaller town that doesn't have the budget to pay for this. Rock stars. Same

Speaker:

with football, right? Whatever sport your thing is, right. You know, for me, I'm a

Speaker:

Yankees fan, although the Yankees have not had a good run of late. But

Speaker:

historically they have been kind of the top. But you know, you

Speaker:

can definitely tell like nation states can be all in like the same league

Speaker:

because they do have more or Less the same capacity in terms of. They're

Speaker:

not, they're not in it for the money per se. Like, you know what I

Speaker:

mean? They're not, you know, they, they. Because they're a

Speaker:

nation state, you know, they can harbor. They can harbor themselves and not

Speaker:

prosecute. You know, they have certain more advantages than your average criminal gang.

Speaker:

Oh yeah. I mean. And well funded. Right. I mean, that's. Money's not

Speaker:

an issue. Yeah, right. It's it. It

Speaker:

puts to be a formative adversary.

Speaker:

And that's why places

Speaker:

like Mandy and that come out with threat reports.

Speaker:

They talk about these actor groups, but you could see moves of

Speaker:

actor groups as well,

Speaker:

changing their tactics and techniques. Again, one of the

Speaker:

more interesting things that happened. I think we could talk about it

Speaker:

because it is public. But if you remember

Speaker:

maybe a year and a half ago, maybe it was two years ago that, that

Speaker:

mgm, you know, casinos guy. Oh yeah, the resort. Remember that?

Speaker:

And it shut. It shut down two casinos with

Speaker:

ransomware. The actors that did that.

Speaker:

It goes back to what's come full circle where it used to be the adversaries,

Speaker:

where it used to be called script kiddies were basically kids

Speaker:

that want to just cause disruption. Well, this is, this is actually a

Speaker:

more sophisticated. It ends up. This group was just a more sophisticated thing of that.

Speaker:

They. Yeah, it was ransomware, but they weren't actually out there

Speaker:

just for the money. They just wanted to do it. They just wanted to see

Speaker:

if they could shut down a casin. And it's crazy that that

Speaker:

that's like that and they still got away with a bunch of crypto

Speaker:

money. But. But it,

Speaker:

you know, it just shows that even like

Speaker:

those, even those hackers could then stand on the shoulders of

Speaker:

all this technology that's being hopefully built for

Speaker:

good and stuff too. But they can use that. And now

Speaker:

you can generate. I know the LLMs.

Speaker:

Back to the subject. You could ask it to

Speaker:

generate malware for you and it'll at

Speaker:

first say no, but if you could trick it, it'll say yes and it'll do

Speaker:

it. And then you can. Didn't that happen recently where there were. State actors

Speaker:

with cl. Yeah, anthropic.

Speaker:

That they were. That. That that plot had been used

Speaker:

and they're, you know, I, I think anthropic really looks at the

Speaker:

safety of what they're doing and stuff too. So they, that's why they disclosed it.

Speaker:

But. And they filled that hole. But it was. It wasn't that hard.

Speaker:

You know, all they did was say, oh no, I'm a Researcher. And

Speaker:

I'm doing an ethical. Yep. It

Speaker:

was not an ethical. How would you. I mean, you've been in the

Speaker:

AppSec before. It was called cyber security. Was called

Speaker:

AppSec or application security. But, you know,

Speaker:

if somebody told you back when you said, you know, no nation state would do

Speaker:

this, right. That, you know, all you had to do is trick a computer into

Speaker:

giving you, like, talk to a computer and tell it you're a research. Like, how

Speaker:

unreal is that? Like, I don't know. I'm doing this for research. Like, oh,

Speaker:

okay. Like, yeah, you know. Yeah, pretty, pretty,

Speaker:

pretty unreal because it was so manual before,

Speaker:

you know where. But again, you know, it gets back

Speaker:

to the thing we talked about earlier. It's like there are builders of the

Speaker:

world that can't imagine someone wanting to destroy,

Speaker:

you know, this beautiful building that's, that's been, that's been

Speaker:

built. And then there's people that. All they think about is, how can I

Speaker:

find a weakness in that building and take it, either take it down or just

Speaker:

gain access and that. Right, that's, that's, that's what's around.

Speaker:

Yeah. I mean, that keeps on. If you're in cyber security, that, that,

Speaker:

that, that keeps the lights on for sure. Because there's always,

Speaker:

always, there's always work to do to help the defenders.

Speaker:

Yeah. So you have something to defend. It was like going back to medieval times,

Speaker:

right. Like, you had the kings, but you had a pretty large class of

Speaker:

knights, you know, that would have to do defending and. Or

Speaker:

I forget what the people were called, but they would stand on the walls and,

Speaker:

like shoot arrows and catapults and stuff like. Yeah.

Speaker:

And you design the moat is because of that. And then

Speaker:

the ways you get into the city, you know, has got

Speaker:

traps in it, you know, and we would

Speaker:

liken that to a honey pot, you know, I mean, there's lots of, lots of.

Speaker:

And then Trojan horse was originally a Trojan horse.

Speaker:

That's right. And it's a battle. Right. And so I think right

Speaker:

now with AI, the attackers have a bit of the upper hand

Speaker:

because we just don't know how they're using it.

Speaker:

But, you know, there'll be tools and there already

Speaker:

are. I mean, if you're a cybersecurity company and you don't have

Speaker:

some AI assistance to help,

Speaker:

either with the scope or the breadth or the speed,

Speaker:

you know, it's. You see that and that. But that's on the detection

Speaker:

end, and sometimes that's too late. I mean,

Speaker:

I hope that the industry moves to some prevention. And then it is

Speaker:

about the building of the moats or the maze that they

Speaker:

have to go through or something like that. And I think that's an important

Speaker:

balance that has to be maintained by enterprises today

Speaker:

to. To make sure that they mitigate the risk. No, that's a

Speaker:

good way to put it.

Speaker:

Any question? We're getting close to the top of the hour, so I want to

Speaker:

be respectful of your time. Any questions? Candace? Sorry, I. No,

Speaker:

honestly, like, this has been a fantastic, fantastic interview.

Speaker:

It's been incredibly enlightening. Like, so much to think about,

Speaker:

you. Know, makes me want to change all my passwords. Right,

Speaker:

right. Well, you know, password 1, 2, 3, nobody. No, you can't. That's not secure

Speaker:

anymore, you know. Well, even, you know, it's interesting

Speaker:

because talk about checkbox compliance versus real

Speaker:

security. Even as we were setting up the company,

Speaker:

little old us, you know, we go to, and we're

Speaker:

needing help because we're wanting to become compliant to some of these

Speaker:

bars that are out there, like SoC2, if you've heard

Speaker:

of that. It's a compliance standard

Speaker:

for trustworthiness of. Of companies like us who might have your day.

Speaker:

There's a massive Alphabet soup there. There is,

Speaker:

there is. And, but like, password policy was really interesting

Speaker:

because what. We were using some AI to help us

Speaker:

in that, and it came back with password policy of. Oh,

Speaker:

yeah, you know, like, change your password every two weeks. Well, that's

Speaker:

been. That might have been state of the art a couple of years ago, but

Speaker:

that's not what you do today. Today, you know, it's about

Speaker:

length and scrambling, and we have these things called password

Speaker:

managers that allow us to do that rather than

Speaker:

us remembering something and. Yeah, even that.

Speaker:

Dynamics. Yes, those can get hacked.

Speaker:

A recent breach on one of those.

Speaker:

You know, there's. I. You know, I don't know, one that was

Speaker:

really bad publicly. There has been in the past, for sure. All right.

Speaker:

Um, yeah, so. But you're still, I think, Correct me if I'm wrong, but I

Speaker:

still think you're safer with a password Manager without

Speaker:

it 100% you. It's.

Speaker:

It's because you want. You want long length,

Speaker:

jumbled, you know, kinds of things that just aren't easy,

Speaker:

easy for the attacker. So.

Speaker:

Yeah, and then you change your master password of that manager

Speaker:

frequently. That's one where you, you. And again, you still want it to be long

Speaker:

and. Right, right, right. Long and complicated. Remember, one long

Speaker:

and complicated thing. Well, I'm, I'm glad

Speaker:

that we got to kind of talk yeah, awesome. It's been

Speaker:

great. Yeah, it's been great. Where can folks find out more about you and your

Speaker:

company? So I think today, you know, like I said, we're

Speaker:

in stealth, but eventually, please follow

Speaker:

Pulse Security AI will be coming out of stealth, you know, in

Speaker:

the, in the probably new year to mid

Speaker:

year kind of thing. But also we started a community

Speaker:

of security professionals and we as

Speaker:

just a networking organization, we call that

Speaker:

securityimpactcircle.org and there

Speaker:

we have blogs. We want to have

Speaker:

people talking about this prevention versus detection or even for

Speaker:

the security leader, about risk and how they should manage

Speaker:

things and best practices that they have together. So

Speaker:

yeah, we have a site, securityimpactcircle.org that is a great place for people

Speaker:

to go and eventually, you know, you'll get to us through that as

Speaker:

well. Cool. Awesome. Well, I'll let

Speaker:

our AI finish the show. And that's a wrap on this

Speaker:

episode of Data Driven. Big thanks to Mike Armistead for

Speaker:

reminding us that while AI may be the future, security breaches

Speaker:

are very much the present. Remember, the attackers only have

Speaker:

to be right once. So maybe don't make your password password

Speaker:

until next time. Stay curious, stay secure, and for

Speaker:

the love of data, please update your firmware. Cheers for

Speaker:

listening. Now go change that password.