Artificial intelligence in backup isn't just marketing hype - it's changing how we protect our data. In this episode, W. Curtis Preston and Prasanna Malaiyandi break down the practical applications of AI in backup systems, from intelligent scheduling to ransomware detection.
Learn how artificial intelligence helps with capacity planning, especially with deduplication systems where predicting storage needs gets tricky. We discuss AI's role in asset discovery, anomaly detection, and even creating better disaster recovery plans. Plus, find out why backing up AI models themselves might become your next big challenge. This no-nonsense look at AI in backup cuts through the confusion and focuses on what really matters - making your backups better.
You've found the backup wrap up your go-to podcast for all things
backup recovery and cyber recovery.
In this episode, we take a look at the use of artificial intelligence in backup.
Can AI make your backup environment actually better?
Prasanna Malaiyandi and I discuss AI and how it can help from
possibly everything from scheduling backups to detecting ransomware.
We talk about using it for deduplication, for capacity planning,
and even helping you to write better disaster recovery plans.
It's time to talk about AI and backups.
Hope you enjoy it.
By the way, if you don't know who I am, I'm w Curtis Preston, AKA, Mr.
Backup, and I've been passionate about backup and recovery for over 30 years.
Ever since I had to tell my boss I. That we had no backups of that really
important database that we had just lost.
I don't want that to happen to you, and that's why I do this podcast.
On this podcast, we turn unappreciated backup admins into Cyber Recovery Heroes.
This is the backup wrap up.
Welcome to the show.
Hi, I'm w Curtis Preston, AKA, Mr. Backup, and I have with me a guy who apparently
doesn't know how to hold a coffee cup.
Prasanna Malaiyandi, how's it going?
Prasanna
I am good, Curtis.
I. So I think we need to clarify a
are you defending yourself?
Are you gonna try to defend your weirdness?
I think we have to talk about multiple things.
First.
In India, they don't typically use like a mug.
They use like a stainless steel cup, right?
So if
you're drinking hot beverages, you can only hold it from like the very,
you saw it when we went to the Indian
restaurant in San Diego,
yeah, yeah.
you have to hold it from the very top, otherwise you'll burn your hand.
Right.
And then most mugs, it just feels weird.
Like I got, I got chunky fingers, like sausages, right?
And so like putting it inside the mug, like the handle part of the mug.
I feel like, especially if it's like a curve, not like a straight, I feel like
there's not enough stability there.
That's fascinating.
So for people watching the video, who, by the
way, we do publish a video on YouTube if you want to see our
glorious faces and our expressions.
But yeah, so when I hold a mug, I don't hold it like this through the
handle.
I basically grab it either from the top or I hold it like on the side.
And then of course, the Pinky's kind
The pinky,
the bottom.
But what's weird though is the pinky supporting the bottom thing.
I know you've complained to me many times, but that's also how I hold my phone.
you end up covering your microphone.
always hold the phone and
then my Pinky's kind of on the bottom, and so it always blocks the microphone.
So Curtis is always
like, were you underwater?
Did you swallow your phone?
What's going on?
So regarding your defense from, you know, how they hold, do things in India.
What part of India were you born in?
Uh, just remind
Yeah, I was, uh, born in not India, but,
but at home.
Right.
But yeah,
you were, you were raised by people born in
India, and so you were, you were taught,
yeah.
And so actually I prefer, so even drinking water.
I don't drink from a glass cup.
I drink from a stainless steel cup.
Right.
Which is, if you haven't spent any time around, you know,
Indians, you wouldn't know that.
It's just that you use a lot, you use stainless steel for cups, for plates,
right.
As Curtis knows what I'm loading, the dishwasher and
he's like, what is that racket?
what is happening over there?
Because everything's so noisy.
They last longer and you don't have to worry about them breaking.
That's, you know, I can't, I can't complain.
Yeah.
Uh, but yeah, I don't get the whole knot, you know?
Here I am with four fingers in my mug.
I'm just saying.
Okay, so now what if that mug was smaller and the handle was curved,
Well, then that's like a, that's like a girly mug and then,
then you use two fingers like
this.
feel like it gives you enough stability?
And yet I've never dropped a mug.
I'm
just saying.
It's not from dropping the mug.
It's from like when you, yeah.
See when you're drinking it, it just feels like it's a little like,
Yeah.
Um,
all over you.
I just think you don't know how to hold a mic, but.
Our listeners are probably like, what are these people talking about?
By the way, this is a new format starting in the new year.
We are now gonna just be talking about coffee and all the crazy
things that Prasanna does.
Yeah, absolutely.
Um, or maybe we might actually talk about some stuff.
So I thought, um, you know, we've been seeing, uh, AI on the news a lot,
right?
ai, I've never heard about it.
Yeah, I've never, never heard of it.
Yeah.
So artificial intelligence, and if, if you've been following the backup
industry much, you probably saw a few announcements from your, uh, backup
company or maybe backup companies you're interested in about the use of ai.
Within backup.
And so I thought we'd talk about that a little bit,
um, in this episode, and
whether or not it has a use, right?
And can, just to clarify, I think when a lot of these backup vendors launched ai,
they were using AI for like the, not for the core product, right?
So they were using AI for their support agent, or to help answer questions, right?
Which I think we all understand, we all know about, but I think in this
episode, I think we should focus on like the core part of backup.
Yeah.
So, so let's talk a just a little bit about, you know,
what we mean when we say ai.
There are different categories of ai and then also there's machine learning, which
is very closely, and honestly, I, I, I,
you know, I think I could describe the difference between machine
learning and ai, but then there's something that, that.
Changes, you know, that, that messes me up when we talk about that.
Um, I'll just, for those of you that actually really know what AI
is and machine learning is, you're gonna be offended by something
I say during this episode.
I, I'll just tell you that.
But we're gonna use the terms almost interchangeably, but they're not.
Uh, but I do want distinguish between.
What is referred to as generative ai, right?
Which is a, you know, a large language model that is
going to create things there.
It's not ex nihilo, right?
It's not from, it's not from nothing.
It's it, it has to, it has to have been trained on a large data set.
But, those are the kinds of things that they're using,
like you talked about there.
Sup for support
models, right?
And, And,
just as examples of large language models, you might've heard about
meta's llama, lama three, Lama four, there's chat, GPT or open ais.
What is it?
OPT?
What,
the, the, actual model.
the
underlying model.
Oh, okay.
I, I, I would just, I would've just said chat, GPT.
'cause everybody knows what chat GPT
is, right?
I mean, you've got copilot, you've got, you've
got, Yeah, you, so you've got Claude from Anthropic.
Um, there are a lot of people, you know, um, confused the company with the product.
But, um, these are the, these are the ones that are grabbing
all the headlines, right?
They're also, they're also writing large bodies of texts.
They're helping people to write books.
They're helping people to do art.
That, and there's a lot of, um.
A lot of legal discussions around that, around the use of things like
the books that I've written as, um, you know, feeding into that and, um,
the, we're not talking about that,
right?
Um, we're not gonna talk about, Hey, um, chat GPT.
My restore didn't work.
Can you recreate all my documents?
Um, it's not,
there's not gonna be anything like that, at least not yet.
Um, the, um, we're gonna talk about how AI can be used to basically
enhance the core functionality.
I mean, you said this in way, a fewer words a few minutes ago,
but, uh, basically how it could be used to make backups better.
And I think a good chunk of this is really, like you said, more
around machine learning models,
right,
right,
large language models.
right.
So the, the first section we will just talk about how potentially just talk about
this is just sort of thoughts out loud.
I know that we have a lot of vendors that listen to the podcast.
We are.
Technically aimed at the, the people who actually use backup and
recovery, but I know a lot of vendors use the podcast, so feel free to
take this episode and run with it and
do stuff.
So I, I guess the first question would be, do we think that, uh, machine learning
can be used to help just to prove the efficiency of the backup process itself?
What do you think about
Oh, a thousand percent.
A billion percent, Curtis.
So I've never actually had to implement a backup system.
But you've done
tons of this, right?
And how do you go about just planning your backup, right?
How to back up an infrastructure, right?
It's like, just walk us through that, right?
And how many spreadsheets and all the rest that you have in
order to try to optimize these.
Yeah, I, I think about that a lot.
And, and, and, and, and the answer is gonna depend greatly on the
product that you're using, right?
You know, I, I can think of.
The traditional way is that you're going to create some kind of schedule, some
kind of, uh, automatic backup schedule.
Um, and you're going to do a, again, traditionally we'll
do three categories here.
Traditionally you've got some full backups and you're gonna do some
full backups every once in a while.
Um, and I was always a proponent if you had to do full backups, I was
always a proponent of doing those.
No.
More often than once a month.
Um, back in the days of tape, it was once a week because
it, was
complicated the restore process.
Yeah.
But, um, you know, doing it no more often than once a month, but depending on your
backup product, you might be able to, to
spread that out even over like three months.
And then you also want to schedule, if your backup product
is capable of doing it, you wanna schedule a cumulative incremental.
A differential, some products call it.
Um, and then of course the daily incremental.
Right.
So spreading
that all
for one application you're talking about,
E exactly.
You're doing this per application, per server.
Um, and, and you're trying to load balance things out because if you've
properly designed your system, it's probably not capable of doing a full
backup of your environment in one night.
Right.
Um, because that would just be really expensive, and then the rest of the
time it would go completely unused.
Right?
Um, so you, you buy it so that it's you, you size it so that it's big
enough to do a full backup over time.
And, um, you're right that, that, that scheduling that out is problematic, right?
Um, and you, you definitely could use, um, uh, AI
or ML to, to do that.
And even for the scheduling aspect.
So we talked about the applications, and then you were talking about sort
of that infrastructure piece, which is shared and you now have to worry
about it across all of these things.
And I'm sure you had these bonkers spreadsheets that you
were creating, trying to do this.
Did it stretch all the way to the moon and back, by the way?
Well, you know me for, it wasn't even a spreadsheet, it was just, uh, it, it was a
script.
Right.
I would, I would just script all this nonsense.
Right?
Um, but it, but it, the bigger the environment, the more.
That doing it programmatically made sense, right?
Um, and, and by the way, even if you have a more modern backup tool
that does incremental forever, there are many applications that
won't, that won't let you do
that.
Right?
I think of like database backups still need to be done every, you know, a full
backup every so often, and you have to schedule these out,
And that's the
second category.
'cause I know you talked about three categories.
Yeah.
Oh yeah.
Oh, well the three categories were, yes.
Uh, thank you.
I'm glad I have you here sometimes, you know.
Yeah.
So you have the, the, the old school full and incremental,
which old school is still current
school.
If we're talking about regular apps, then there's the forever incremental type.
Um, and you don't, you, you do have to worry about scheduling those,
but generally you just sort of tell 'em all to start at once and then
they queue and then it is not, it's, it's a lot simpler to do those.
I. But then the final category are ones that actually, um, and I
think the one that probably stands out the most here would be Rubrik,
right?
Rubrik doesn't let you schedule, um, that
stuff.
You tell it what your RTO
is and your RPO, and it just does the backups.
I mean, in fact, there are people that complain that you cannot, at least
last time I checked, you could not do.
a a manually scheduled backup if you wanted to tell it when to do stuff.
Um, I, I think this is probably the first use of some sort of machine learning
or artificial intelligence that I can think of with regards to scheduling.
Which, which I was also gonna chime in.
So the first two methods you talked about, right?
You're kind of statically doing this upfront, setting the schedules and
hoping that forever that it will be good,
Right.
You'll always be able to meet it, but say that there's an additional load or a
server goes down or something else, right.
There's no way to fine tune and adjust that,
Well, well, I, Well, there, I mean, there is, but there's
no way to automatically fine
tune and Yeah.
Yeah.
Right.
And so you're just like, okay, maybe it'll fail a couple times
and then I'll adjust the policies and then I'll be fine, but Right.
Versus something like an SLA based, which I, I actually have
looked at rubrics in the past,
and I find that very enticing because really in the end, you
care about what your RPO and RTO,
Yeah.
No one cares if you can back up.
They only care if you can restore.
the problem though is it's such a big paradigm shift for a lot of backup admins
that it's very difficult to understand because it's like when people move
from on-premises to the cloud and they were concerned because they're like,
I can't touch and feel my equipment.
Right.
It's not something I could actually do.
I think that's also the same challenges you get when you move
from sort of, uh, schedule-based backups to sort of SLA based backups.
Yeah, I, I liked, I liked the idea a lot.
I, I, I still, again, you know, if I was, if I was running rubric,
I would give people the ability to do a manual backup if they
wanted to.
But, but I do really like the idea of SLA driven backups,
because I like the idea of SLAs.
You know, we've talked about SLAs on here, and I like the idea of.
Knowing the back backups were being done often enough to meet my SLAs.
I
really liked that idea.
The one thing I think that is useful with these sort of approaches is
we've talked about the fact that like your environment doesn't say static.
Right.
So as you're adding new workloads, as things are changing, you don't
want to have to go recompute your entire spreadsheet or your
script H every single time.
So it's nice to have sort of these models that can automatically help fine tune and
optimize so you're not wasting your time because it's more than likely that you're
not gonna get it right the first time if you manually try to reset some of these
things.
And so having this automatic thing that constantly is
adjusting just seems amazing.
Yeah, it does.
And I, and outside of Rubrik, I'm not aware of any tools that do that.
Uh, but I, I think that this could certainly be a way where
they could use AI to do that.
Um, the.
And I, and I was thinking about, again, going back to it, it's been a
while since I've had to do this in a production environment, but the, the
the first thing that you have to find out is how big is everything, right?
How big is, is everything from a database perspective and
how, how long does it take?
'cause there's all these different, and that's the thing that nobody knows.
Right.
How big is your, how big is your data center?
And they're like, I don't know.
I don't know.
And so like, you have to do a full backup first
before you have any idea.
And not every server backs up at the same speed and all these different things.
So yeah, it it is a
complicated
and you may not be able to back up everything at the same
time because there might be
different hours, right?
That
a server is sort of offline or has less load that you can actually do it.
Yeah, so having some sort of AI or ml, um, figure that out sounds amazing.
Right?
Another area where I think that this could help is very, very closely related, and
that is, and, and some backup products do have this and that is making sure
that everything in my data center.
Is backed up in some
way, right?
Usually where you see this is an integration with like, um, uh,
VMware or, uh, AWS, et cetera, right?
Um, basically just connect to my entire, uh, you know, control
panel and then just look and make sure that everything is connected
to some type of policy to back it
up.
I, I think.
a default policy if anything is created, so at least everything
is protected, even though
it may not be protected with the right thing, but at least it's
being protected and you don't have to worry about these gaps.
I.
Exactly.
Exactly.
Um, and I, I think you do see this in a lot of backup products.
Usually again, it's with integration
with, uh, big things like VMware, HyperV, AWS, um,
you know, et cetera.
you need the companies, those vendors, to actually provide the APIs to be
able to do these sort of queries, and I think that's where there's kind
of a little bit of a tension there,
Yeah.
Yeah.
I mean, theoretically you could scour the data center, right?
Uh, looking for new computers.
Again, I, I know I mentioned this before, but you know, back
in the day we did that, right?
And back in the day we did that with Vizio.
Um, the, the vis, there used to be a very
expensive version of Vizio that would just literally crawl your data center.
And it used, uh, some very interesting technology.
Um, I forgot the, the name of this, but like, inmap
does this, where it, what it does is it sends a malformed packet.
It finds an IP address, it sends a malformed packet to that IP address
to see how it responds, and different things respond in different ways.
And that's how it, that's how it, um,
That
is crazy that they built that.
Yeah.
Yeah.
Um, and so you, you could theoretically do that, but a agreed, it's much easier
if you just have, everything's gonna be in VMware or AWS and then just talk to AWS.
Now again, going to VMware and AWS, there can be multiple virtual data centers.
There can be
multiple AWS accounts.
So you, you, you want to make sure that, that you have some way to, to
do that.
And I, and I do like that idea.
Shadow it.
Yeah, shadow it bad,
especially when it comes to backup.
Right.
Um, again, I'll tell a story from back in the day was the time that someone came to
me and they had, they were DBAs and they, they gave me a directory of a database.
They wanted me to restore.
Restore, and it was temp, um slash TMP on a, on a HP box.
And for those that don't know slash TMP on an HP box specifically, HPUX was in ram.
So when you rebooted it, temp went away.
And this, um,
this
it source code,
what I.
it?
Source code
It was source code.
Yeah.
And they were developing for months, like an entire team of
developers developing source code of this new application in temp.
And then we rebooted the server and they, and they came to me
and asked me to restore it.
And I was like, dude, we don't back up temp. I don't know
what you're talking about.
Like, and they're like, dude, this is really important,
like heads are gonna roll.
And I'm like, yeah, not mine.
Like everybody knows we don't back up temp.
Except for you, apparently.
Oh
Uh, so it's, I'm just, you know, it's really bad when you have
a functioning system and then it's not being backed up again.
Another story we used to have, um, we had a, a naming convention.
Ours was very boring.
Um, it, it was, it H-P-D-B-S-V-A, right?
HP database server A, and there was HB FS oh one, et
cetera, right?
And I remember, and I had this form that you had to fill out.
This was an actual piece of paper.
We did
not have web pages.
Right?
You had this form that you fill out and, and you had to, and, and it, it said
on there, simply filling out this form is not, does not meet the requirement.
You do not consider your system backed up until you have a signed form back from me.
Right?
And then one day somebody handed me a form and it said like.
They wanted, like me to back up H-P-D-B-S-V-M, right?
And I go, M that's interesting.
The last server I remember hearing about was H. So that means there's an I, A
J, A K, and an L out there somewhere.
hasn't been backed up.
That hasn't been backed up.
Yeah.
Um, so this idea of automatically
detecting servers and applications sounds like a great
idea.
And also not just VMs, but also detect, it would be really
nice if it detected the type of
VM and said, this appears to be a SQL instance.
We should back it up with the default SQL
policy.
That would be great.
So in addition to making things more efficient, um, there are some
other things we could do, uh, with AI that also would be interesting.
Uh, what do
you think is the, the first one?
No.
So I think one of the ones, and we've talked about it so much, so often,
and vendors are starting to do this, it's around anomaly detection and
it could be used in various fashion.
So one thing is like, Hey, by the way, this server, all of a sudden it's backing
up 10 times what it normally does.
Maybe this might indicate like a malware or ransomware on the system.
Right?
Um.
Or Hey, I've noticed that there's a bunch of data that's starting
to look like based on entropy.
That it's been encrypted, that doesn't look normal.
Okay, maybe I should go investigate it, right?
So, or it could even be security things like, Hey, you're logging
in from a different place than normal as a backup admin.
Is this the right thing or not?
Yeah.
And also very closely related to the stuff you said before was, uh,
are files where the file type based on the first few bytes of the file,
does not match the extension of the
file.
So it says it's a dot doc, but the first few bites of the file
show that it's an application, for
Sorry, one
Yeah, that's an interesting use case around, uh, the first few bites because
that could detect things that are being encrypted or other things that don't
make sense, or potentially even malware.
Right.
Yeah, it, uh, it's something we do, you know, my, uh, employee is S two
data and we do a lot of restores of old stuff, um, where we're pulling
data off of tape often for, um, I. For e-discovery purposes and lawsuit
purposes and, um, investigation purposes.
And one of the things that we do as we're pulling data, 'cause we
use a, a, a proprietary tool that we've written to restore data off
of most backups rather than use the built in tool for a lot of reasons.
Um, and this is one of them is that we check the file type against the file
contents and, uh, it can, it can also indicate.
Um, uh, subterfuge,
right?
Um, it can indicate somebody trying to hide something.
Um, but yeah, so anomaly detection, I think is a really big one.
Uh, right.
Definitely that this is a, this is a, you looks like you've got ransomware, right?
You need
to solve that.
That was probably the, the first big use of AI that I
remember, uh, in, in the backup world.
And I, I, I will say that if.
The way that you know, that you have ransomware is that your backup
product told you something is wrong, but, uh, but it, but it can
happen.
Right.
Um, another one that I'll talk, uh, that I'd bring up is, is data classification.
Again, I think that.
This is, this is probably a very simple one, but the
idea of like, looking at all the different data types and helping you to
understand what is in your environment.
This is not that new.
Um, but perhaps the AI use case could be helping you to identify trends,
um, and, and where the data's moving, where it's being created, where
it's being changed, uh, et cetera.
Um, and, and then, which is very closely related to my
other idea, which is predictive
analytics.
Right.
Um, again, going back to, uh, you know, back in the day,
one of the things I remember being the hardest to do is capacity prediction.
You
know, predicting whether or not I have enough capacity To
do my backups for the next six
and you know what makes it even harder?
What's that?
It does, d ddu makes it way harder.
And you know what AI right?
Ai ml could, could use to, could be used because it's smarter than I am.
Smarter than you are.
It could actually understand the trends
as to now what, what, let's talk about that Non, not every,
everybody might not understand.
Why DDU makes capacity,
Sure.
uh, management so
So let's talk about the, before we get to D Dub, let's talk about like
traditional storage or tape, right?
So
you're doing a full backup, you know how big your database is, therefore,
you know, okay, my full backup is gonna take this much space and
you know, with compression, maybe it's gonna be two x or half the space, right?
And then, you know, okay, my daily change rate is say 5%, and based on the
total size, I know what that's gonna be.
And so
if I'm doing weekly fulls, daily incrementals, I know how much
storage I'm gonna need for a week.
Yeah.
And, and just as, and just as important, you also know how
much storage, when you delete
the, you know, the older backups.
Yeah.
You know how much storage will be freed up, which is just if, if not even more
important.
Now the problem with deduplication is they talk about these great rates like
40 x, 30 x, 20 x, take your pick, right?
And that's all great.
If you're all like if a lot of your data is very similar, but it's hard
to tell, is your data similar or not until you've actually start doing it.
So if you're trying to buy storage for, say, three years
ahead of time, a capacity plan.
It becomes really difficult.
And so you guess, right?
You'll take a stab and maybe you look at some of your data and you're like,
Hey, these kind of look the same, but you don't know if that's right or not
until you actually start backing it up.
And like you said, Curtis, if you go delete your backup, you may not
actually free up that space because it's been de-duplicated against something
else that you're still preserving.
right,
Say I go delete my backup for six months ago for one application.
Another application might have, uh, common blocks with that data or with that other
application.
And so even though I deleted the first application's backup,
it's not gonna free up space.
And so you end up with this problem and this challenge.
And that's one of the things, the hardest things about deduplication.
Having worked at a company that did deduplication, customers
always struggled with it,
Yeah,
And some of the
things we would do is we would be like, Hey, let's scan your
application and just understand what sort of DDU rates you may get.
And even that's a guess, because maybe you move an application from one storage
appliance to a different appliance and now your DDU rates are different.
Yeah.
And, and, and again, the
one of the most frustrating things could be if you, you start.
You're running outta capacity, right?
And so you say, listen, I know we said we wanted to keep backups for
three years, but we're running outta capacity and so we're gonna start
deleting three years minus a month.
And you do that and you get
back 0.1% of your, it can be very difficult.
Um,
fact that to free up that space takes time.
Because typically with a lot of these systems, there's a background process
typically called garbage collection,
which goes and now needs to free up all this data and that does take time to run.
Yeah, it is, it is a two stage process where you, you, you, um, flag that
block for deletion and then another
process that runs typically when backups aren't running.
Um, and you, you probably have to force the garbage collection process.
Um, so go, go ahead.
so I was just thinking as we were talking about the first time
that I heard about AI in storage,
and I think the first company that I can recall, and I'm sure there
were others, was actually nimble.
Storage and nimble.
What they did is their first product when they built they, so
they provided primary storage.
And their first product, they basically were like, Hey, we are optimized for sql.
We are optimized for VMware.
We are optimized for these different, and I was like, oh, that's pretty awesome.
They're doing it dynamically.
But I think at the time it was kind of a static thing where you
would say, Hey, I have VMware.
I'm writing into this data store.
And it would optimize its, and it would basically pick different
block sizes for deduplication
Right, right, right.
Yeah.
That's interesting.
The, the, the, I, I, I think div, going back to the thing
we were talking about of like.
Using AI to basically help me understand when do I need to order more storage?
It can, to the best of its ability.
It can actually look at all of the DDU rates, right?
At all of the at, at what?
It could look at the DDU rate of each individual backup, right?
You, you gave, you told me it's a backup this much and this is
how much, and so we can actually
run all those calculations and I can actually figure out.
Well in six months, based on if everything stays the
same in six months, you're gonna be
outta storage.
So
many vendors actually do.
Yeah.
Yeah.
Um, so the, the, um,
Because I think storage capacity is a little easier.
To predict, because like you said, you're not really changing things, right.
You know what your policy is.
You know what data's coming in, you know how long it's, you're keeping it,
you know what your deduplication rates are, you know how much it's filling up.
So I think it's a little easier than what we had talked about previously
where it's like, okay, now let me plan out my entire backup infrastructure
and start scheduling that.
Yeah.
Speaking of dedupe, can AI help dedupe itself?
Do you think that?
can.
So I think my biggest.
Challenge would be that to run AI requires compute
and usually backup.
You want to go as fast as you can,
Mm-hmm.
right?
And so I think there's that tension.
That exists between running as fast as you can versus introducing
something in the pipeline to that could potentially slow things down.
And you'd have to also ask at what cost, right?
Like, are you going to be saving, say 70% additional versus a traditional
algorithms, or is it gonna be much less
Yeah, I think ddu in, um, in the backup world, there, there, there
have been two main ways to do ddu, which has been, there has been
something that isn't really ddu, but
there were DDU products that called themselves DDU products that did this.
Uh, and that would be block level, um,
incremental, essentially.
Right?
Not
actually de-duping things against each other, but just.
Using technology to lower the additional new data that's
backed up from each workload.
But then the traditional ddu, the way it works for those that don't know
this, is that you slice it up, you slice everything up into what are
typically called shards or chunks.
You run some type of algorithm on it that gives you some type of thing.
Like, like
A fingerprint.
the original SHA two,
SHA 2 56.
And again, here the, the better the algorithm, um, the better the ddu,
but the better the algorithm, the more compute it takes going back to
your trade off thing.
And so, um, that's the way basically every chunk it's run through, you come
up with this alpha numeric string, that alpha numeric string is compared
with every other alpha numeric string.
I. Um, and then that's how you identify redundant data.
And one of the challenges you have with that method is that, uh, the data slides,
um, and so if you don't slice the data at exactly the same spot it, it's duplicate
data, but you don't, don't identify it.
The, there is a completely different way which, um, you
look at the way vast does things.
They do something completely different, right?
So they, they have an algorithm and, and I, I'm guessing they
use AI or ML to, to, do this.
They have an algorithm that, um, basically identifies data that
is probably redundant, right?
Um, that, that, so they, they've got two different ways to do de-dupe and I, so
there are potentially, again, potentially.
AI or ML could be used to identify a new way to identify duplicate
data that is maybe, maybe
more efficient from a compute and storage.
Like even if it was just more efficient from a compute standpoint,
but got the but got the same amount of dedupe, that would still
be great.
Um, but
potentially this is something
that I think, uh, AI could
and the one thing I did also want to comment on Curtis is, uh, going back to
your comment about, okay, if the data shifts, then now you have to make sure
that you're doing the right blocks, right?
Uh, this is where companies though have done sort of, uh, what you're
talking about is called fixed block.
Fixed block deduplication,
right?
There are
many vendors out there though, who do variable size.
Variable block, uh, deduplication, which allows it to vary such that if
you do get an offset right, because of some data change, it's still able to
dup everything else after that because
of how it's actually computing the chunks, the segments, right?
Each of
the blocks.
Yep.
Um, so, uh, so that, that's certainly an area where, where AI could potentially
help the, um, the next, do you think it could help with recovery testing?
Oh yeah, I would.
So one thing for C is like, most people probably don't
know how to write a DR plan,
Mm-hmm.
Mm-hmm.
right.
Um, I wonder if you took ai, like even, and I'm going back to the first
set, right, the large language models,
Yep.
So the thing we said we
weren't talking about, I think we're gonna talk about it here.
Yeah.
I think at least to start with, it's like, Hey, here's all my data.
Here's my applications.
Help me build a DR test plan.
Yeah,
I like that idea.
And
see what it pops out because, and it may not be perfect, and don't just
blindly trust what it provides, but use it as a starting point, right?
And then go use that.
Because I think a lot of people struggle with, where do I even start?
Yeah.
And you could also, um, you could use it like a chaos monkey,
right?
You could use it.
Help me come up with some interesting scenarios.
To just make the, the idea, you know, one of the things that we talked about with in
terms of, uh, cyber testing, uh, was, um.
You know, when we had Mike on the idea of like, doing this and, and
making it, making it fun, making it a game, uh, I like that idea a
lot and I think maybe AI could help
there.
Um,
if, if it helps you do recovery testing more often, um, and, uh,
helps you identify potential, uh, uh, plot, I was gonna say plot
holes, uh, potential, potential holes in your program, uh, then that, then that
I think could be, um, very
helpful.
And Curtis, since you threw out a term, Chaos Monkey is a tool that was released
by Netflix, and literally what it is used for is to just test it, resiliency.
So it'll go randomly, kill services, kill locations, kill
network connections, just to see.
Is streaming, interrupted, are, uh, end users having any sort of
issues and it's able to do this at a scale and in an automated fashion
versus someone like trying to think about all the combinations,
permutations, and scenarios, because they're probably gonna miss things.
And so Netflix designed this thing to actually go out and
test their infrastructure.
It is pretty impressive.
Uh, you know, their infrastructure in general is pretty impressive.
It's not flawless.
Um, I did, I did watch part of the, uh.
The Tyson fight a little while ago, and that was on Netflix
and it was not good, right?
That wasn't so much a resilient thing as it was.
They just, again, they could have used perhaps a little bit better
AI to predict the, what kind of load they were gonna have.
But yeah.
But the idea of predicting crazy things that will happen, uh, Netflix
is pretty darn resilient, uh, when it comes to their infrastructure,
Yep.
yeah, I, I like that idea a lot.
Um, and, and I think, I think this is something that could be, that,
that, that, again, an, uh, uh, an LLM could actually help with, right?
So, like I said, the thing that we said we weren't gonna talk about,
we could talk about it, right?
Um, and for those, if you've never used a chat, g PT or a Claude,
uh, I think it's very useful
here, right?
You, you could say, Hey, I, I'm this kind of company.
This is the type of company, you know, and I understand the,
the privacy concerns of what you
share with a chat g pt or a clot.
Uh, there, there are, by the way, there are on-prem versions that
you can run, uh, of these LLMs too, so that you can keep the
data to yourself.
But the, you have a conversation with it.
Here's the type of company I am, here's the type of computing environment I have.
What do you th what could go
wrong?
Um, you know what, what could I build a, a dr scenario
around?
Any final thoughts?
Can you think of, uh, any other areas where we could use AI and, and backup?
Not so much.
I think the one thing I do wanna call out though is AI is here to stay.
ML is here to stay.
Don't be afraid of it.
Use it.
Right in the right ways and don't be afraid and just start thinking about it.
Uh, the one other thing I will call out is as companies are starting
to dig into AI and ML for their own applications, production applications
and other things, as a backup admin, you need to start thinking
about how do I protect this, right?
How do I back it up?
How would I potentially restore it?
Because there's a lot of data and training these models.
Is really, really expensive.
Mm.
And so you wanna make sure you have mechanisms to protect the models
that emerge from all of this training so you can restore them if needed.
So use backup to, to make AI more resilient while AI makes backup more
resilient.
I like that.
We'll call that a symbiosis.
I like that a lot.
Uh, one my final thought is that potentially you could use, again,
going back to the thing we said we weren't gonna talk about.
You could use LLMs to help select vendors, right?
You could say, Hey, here are all my requirements and here's all the
documents that they, they gave me this 57 page response to my 10 page RFI.
Can you help me make sense of it?
Um, and, uh, you, you could use that again, trust but
verify when using an LLM for
sure.
All right, well, thanks again, Prasanna, uh, for a good chat.
Thank you, Curtis.
And I am not gonna change how I hold a coffee mug.
I'm sorry.
I, I would expect no less.
And thanks to our listeners, uh, we'd be nothing without you.
That is a wrap.