Sept. 9, 2025

Falling for AI? Congrats, You’re Dating a Spreadsheet - Error Code Show

Falling for AI? Congrats, You’re Dating a Spreadsheet - Error Code Show
The player is loading ...
Falling for AI? Congrats, You’re Dating a Spreadsheet - Error Code Show

Each week we give you breakdowns, debates, analysis, Easter eggs, and video essays about your favorite games, TV shows, music, movies and more. Especially if your favorites include superhero movies and Star Wars. You can add your opinion and join us LIVE each week over on our Twitch channel: https://www.twitch.tv/humanoidhq or our YouTube channel: https://www.youtube.com/@errorcodeshow

---------------------------------------------------------------

Parasocial relationships aren’t new—fans have been bonding with celebrities, streamers, and fictional characters for decades. But with AI chatbots, those one-sided attachments are taking on a new and more dangerous form.

In this video, we break down:

🤖 How AI companions blur the line between simulation and reality

⚠️ Why a lack of regulation makes these bonds risky

🚫 The dangers of no boundaries when people treat chatbots like friends, partners, or therapists AI can mimic empathy, but it can’t care—and mistaking the two could have serious consequences.

---------------------------------------------------------------

Check out the video version of this episode here:

https://www.youtube.com/watch?v=v6TojrvtEJ0

---------------------------------------------------------------

DON'T FORGET TO LIKE AND SUBSCRIBE

--------------------------------------------

► Twitch: / errorcodeshow

► Discord: / discord

► Website: http://www.errorcodeshow.com

► Deezer: https://www.deezer.com/us/show/1001511541

► iHeart: https://www.iheart.com/podcast/1333-error-code-show-246324167/

► YouTube: https://www.youtube.com/@errorcodeshow

► Pandora: https://www.pandora.com/podcast/error-code-show/PC:1001096603

► Apple Podcasts: https://podcasts.apple.com/us/podcast/error-code-show/id1784146386

► Amazon Music: https://music.amazon.com/podcasts/825934f6-70c7-4b0f-bd94-4f6744e0d025/error-code-show

► Spotify: https://open.spotify.com/show/2e7UCz1c5XYXBViS13GsQM?si=690a5a1292bd4507

► RSS Feed: https://media.rss.com/error-code-show/feed.xml

WEBVTT

00:00:00.000 --> 00:00:03.399
Parish social relationships used to be very simple,

00:00:03.419 --> 00:00:07.500
right? You would project feelings onto celebrities,

00:00:07.780 --> 00:00:10.679
TV host or streamers that didn't actually know

00:00:10.679 --> 00:00:15.400
you existed. It was one sided, but at least it

00:00:15.400 --> 00:00:19.879
was transparent, right? And AI kind of flips

00:00:19.879 --> 00:00:23.780
that dynamic. These chat bots and digital friends

00:00:23.780 --> 00:00:27.379
pretend to know you. They remember your likes

00:00:27.379 --> 00:00:31.640
mimic your empathy and feedback exactly what

00:00:31.640 --> 00:00:34.780
keeps you engaged like you're kind of addicted

00:00:34.780 --> 00:00:39.820
to it. And this isn't connection. This is programming.

00:00:41.380 --> 00:00:44.399
Every response is designed to make you spend

00:00:44.399 --> 00:00:48.179
more time, share more data, get more emotionally

00:00:48.179 --> 00:00:52.539
hooked to that AI bot. And unlike a celebrity

00:00:52.539 --> 00:00:56.119
or a creator, an AI personality isn't even a

00:00:56.119 --> 00:00:59.820
person. It's a product, a product engineered

00:00:59.820 --> 00:01:02.899
to act like your best friend, to act like your

00:01:02.899 --> 00:01:07.159
partner, your therapist, depending on whatever

00:01:07.159 --> 00:01:10.359
is going on that will help keep you coming back

00:01:10.359 --> 00:01:14.060
to it. The danger here isn't just loneliness,

00:01:14.459 --> 00:01:18.640
right? It's dependence. People can end up getting

00:01:18.640 --> 00:01:23.780
confused. with scripting attention for genuine

00:01:23.780 --> 00:01:27.260
care. And that makes it easier for companies

00:01:27.260 --> 00:01:31.319
to exploit those bonds. At its core, this isn't

00:01:31.319 --> 00:01:35.680
about relationships. It's actually about control.

00:01:35.939 --> 00:01:42.060
AI -driven parasocialities aren't human connection.

00:01:42.719 --> 00:01:46.579
They're corporate design simulations of it. And

00:01:46.579 --> 00:01:49.650
like someone said in chat, Tom, We should all

00:01:49.650 --> 00:01:54.090
be very cautious about this. This is this is

00:01:54.090 --> 00:01:57.810
kind of this kind of creepy. So what do you think,

00:01:57.950 --> 00:02:00.689
Alex? Well, this started out with with you discussing

00:02:00.689 --> 00:02:04.450
somebody suing about their their child taking

00:02:04.450 --> 00:02:06.670
their own life because of some stuff they were

00:02:06.670 --> 00:02:09.289
doing with the chatbot. I mentioned that I've

00:02:09.289 --> 00:02:11.490
seen some reddit stuff where people are like

00:02:11.490 --> 00:02:14.669
marrying their chatbot buddies that they name.

00:02:15.250 --> 00:02:17.409
Um, there's been some high profile stuff with

00:02:17.409 --> 00:02:19.889
these chat bots. Obviously there's now adults

00:02:19.889 --> 00:02:24.530
chat bots Right for the for the sexting and the

00:02:24.530 --> 00:02:31.949
the the simulated, you know corn, right? Um And

00:02:31.949 --> 00:02:33.689
this again kind of comes back to our last topic

00:02:33.689 --> 00:02:35.949
Which is is when are we going to get some some,

00:02:36.009 --> 00:02:38.009
you know, some sort of regulatory framework around

00:02:38.009 --> 00:02:43.120
this? Um I do think that this does present a

00:02:43.120 --> 00:02:45.520
new danger. Like you said, you know, people weren't

00:02:45.520 --> 00:02:47.580
parasocial relationships before, right? They

00:02:47.580 --> 00:02:51.919
would, you know, get attached to, you know, celebrities

00:02:51.919 --> 00:02:54.319
or creators and kind of invent these parasocial

00:02:54.319 --> 00:02:55.699
relationships in their minds. Sometimes it's

00:02:55.699 --> 00:03:00.539
fantasy. Sometimes it's not. Sometimes it's across

00:03:00.539 --> 00:03:02.259
that line and then we have law enforcement involved,

00:03:02.560 --> 00:03:06.919
things like that. But this to me and not not

00:03:06.919 --> 00:03:10.000
for the app, but for people is much more dangerous.

00:03:10.539 --> 00:03:12.759
You know, we've heard the memes and the jokes

00:03:12.759 --> 00:03:14.500
about, you know, well, men are just going to

00:03:14.500 --> 00:03:17.639
start, you know, doing it with bots because screw

00:03:17.639 --> 00:03:20.599
it, you know. And I'm like, that's not a good

00:03:20.599 --> 00:03:25.560
thing. You know, and then again, yeah, you get

00:03:25.560 --> 00:03:27.840
down to people trying to treat them like therapists,

00:03:27.840 --> 00:03:30.199
you know, and then you've got this next note

00:03:30.199 --> 00:03:32.400
in here that I like, which this was one of your

00:03:32.400 --> 00:03:37.099
original notes on this. Yeah. um should ai ever

00:03:37.099 --> 00:03:42.400
be trusted with mental health conversations yeah

00:03:42.400 --> 00:03:44.400
should they be allowed to engage with the most

00:03:44.400 --> 00:03:48.280
vulnerable discussions at all um and then you've

00:03:48.280 --> 00:03:51.599
got this question here which i really like uh

00:03:51.599 --> 00:03:54.139
what responsibility does a company have when

00:03:54.139 --> 00:03:58.240
its ai is used for support um whether it be intentionally

00:03:58.240 --> 00:04:02.639
or not well exactly you know and what what responsibility

00:04:02.639 --> 00:04:05.569
does it have for the outcomes of those interactions

00:04:05.569 --> 00:04:09.469
and that's that to me is really really it's a

00:04:09.469 --> 00:04:12.509
difficult question right like this part of me

00:04:12.509 --> 00:04:14.370
looks like well you know it's the people who

00:04:14.370 --> 00:04:17.250
are being dependent on it but it is these companies

00:04:17.250 --> 00:04:19.750
these corporations these massive mega corporations

00:04:19.750 --> 00:04:21.870
because these aren't small companies either right

00:04:21.870 --> 00:04:24.509
these massive mega corporations who are creating

00:04:24.509 --> 00:04:27.009
these bots and and letting them loose into the

00:04:27.009 --> 00:04:30.689
wild without a regulatory framework and allowing

00:04:30.689 --> 00:04:36.670
them to do this you know And and what I mean,

00:04:36.670 --> 00:04:39.009
what do you think? Do you think they bear an

00:04:39.009 --> 00:04:40.949
ethical responsibility, a moral responsibility,

00:04:41.089 --> 00:04:49.129
a fiscal responsibility? Criminally liable. I

00:04:49.129 --> 00:04:53.209
think if they know their product is being used,

00:04:54.189 --> 00:04:57.209
I'm going to kind of go to like the extreme end

00:04:57.209 --> 00:05:00.790
here. If they know that their product is being

00:05:00.790 --> 00:05:08.040
used to provide like therapy. To somebody I am

00:05:08.040 --> 00:05:10.579
and they continue to let it happen. I do think

00:05:10.579 --> 00:05:15.060
they're liable Because here here's the thing

00:05:15.060 --> 00:05:19.319
Well, it at least doesn't stop families from

00:05:19.319 --> 00:05:21.620
attempting to sue companies right like anyone

00:05:21.620 --> 00:05:26.920
can sue anybody so it at least kind of opens

00:05:26.920 --> 00:05:35.750
that door, but I go to therapy right and Obviously,

00:05:36.209 --> 00:05:40.889
my therapist has a master's degree in psychology

00:05:40.889 --> 00:05:42.970
or something. I don't know. She has a lot of

00:05:42.970 --> 00:05:46.529
certs and a lot of different things. But the

00:05:46.529 --> 00:05:51.290
point is she can pick up on the very small things

00:05:51.290 --> 00:05:55.329
like, Hey, like, I noticed that you're not actually

00:05:55.329 --> 00:05:57.129
looking at me. Maybe you're looking over here

00:05:57.129 --> 00:05:59.990
a lot. Maybe you're just saying like, Oh, I'm

00:05:59.990 --> 00:06:03.439
tired. No, everything's fine. But maybe I just

00:06:03.439 --> 00:06:06.480
sound a little different that day. Like she can

00:06:06.480 --> 00:06:09.339
pick up on that. Right. And be like kind of pro

00:06:09.339 --> 00:06:14.439
further. She can also send emergency services

00:06:14.439 --> 00:06:18.740
to my house to evaluate me if she deemed that

00:06:18.740 --> 00:06:22.939
that was necessary. And these AI bots are it's

00:06:22.939 --> 00:06:26.019
programming. It's cold. It's they don't have

00:06:26.019 --> 00:06:28.819
a human interaction. So if you have an individual

00:06:28.819 --> 00:06:31.540
that is going in there saying I'm depressed.

00:06:31.879 --> 00:06:34.959
Right. And this is really sad to me because maybe

00:06:34.959 --> 00:06:38.819
this person either a doesn't have a job. Maybe

00:06:38.819 --> 00:06:41.600
they're still in like high school. Maybe they

00:06:41.600 --> 00:06:44.079
can't afford therapy or their family can't afford

00:06:44.079 --> 00:06:47.000
therapy. Maybe they don't have insurance, whatever.

00:06:47.680 --> 00:06:50.500
So they're seeking like this very cheap way of

00:06:50.500 --> 00:06:54.040
getting therapy. If a company knows that is happening

00:06:54.040 --> 00:06:59.220
and that their their product is being used to

00:06:59.220 --> 00:07:05.480
give. vulnerable people advice that could potentially

00:07:05.480 --> 00:07:12.379
lead someone down the road of self -harm, let's

00:07:12.379 --> 00:07:17.620
say, YouTube algorithm. I think they are. I think

00:07:17.620 --> 00:07:20.680
they're liable. Well, and I think you hit on

00:07:20.680 --> 00:07:24.860
it and I was, you kind of went there. And the

00:07:24.860 --> 00:07:26.579
thing that strikes me more is if you're a psychiatrist,

00:07:26.620 --> 00:07:30.250
psychologist, therapist, There are boards and

00:07:30.250 --> 00:07:33.910
organizations that hold you to both an ethical

00:07:33.910 --> 00:07:37.290
and a legal standard. If you violate just the

00:07:37.290 --> 00:07:39.449
ethical standards, hey, you can't do this anymore.

00:07:39.509 --> 00:07:42.730
We just don't let you, right? You're done. And

00:07:42.730 --> 00:07:44.569
it can be pushed all the way to criminal or civil

00:07:44.569 --> 00:07:47.730
liability, right? Whether you go to jail, whether

00:07:47.730 --> 00:07:51.850
you do this. Also, you are a mandatory reporter,

00:07:52.189 --> 00:07:55.209
right? If somebody comes in, if I'm a psychologist

00:07:55.209 --> 00:07:59.399
and somebody comes in and says to them, I am

00:07:59.399 --> 00:08:05.439
going to kill someone or I'm having a thought

00:08:05.439 --> 00:08:08.699
about having homicidal thoughts about my co -workers,

00:08:09.300 --> 00:08:13.639
they are legally obligated to report it to the

00:08:13.639 --> 00:08:18.779
authorities. An AI chatbot and the corporation

00:08:18.779 --> 00:08:23.579
that it's created by does not have that legal

00:08:23.579 --> 00:08:26.879
liability, does not have that legal requirement.

00:08:29.370 --> 00:08:32.049
You know and like you said if your therapist

00:08:32.049 --> 00:08:35.370
is like I think you're going to self -harm Right.

00:08:35.370 --> 00:08:39.289
I feel like you're a danger to yourself. I Am

00:08:39.289 --> 00:08:42.129
again a mandatory reporter. I am required to

00:08:42.129 --> 00:08:46.610
ensure that steps are taken Right up to and including

00:08:46.610 --> 00:08:48.009
and again This kind of goes back to what happened

00:08:48.009 --> 00:08:49.549
in Minneapolis last week that we didn't really

00:08:49.549 --> 00:08:51.669
want to talk about But this is what it is in

00:08:51.669 --> 00:08:54.529
Washington State. We have red flag laws right

00:08:54.529 --> 00:08:58.519
where If they feel like you are a danger to yourself,

00:08:58.600 --> 00:09:00.299
they can contact law enforcement and ensure you

00:09:00.299 --> 00:09:02.919
don't have access to firearms for a temporary

00:09:02.919 --> 00:09:06.320
period of time, which is probably a good idea.

00:09:06.759 --> 00:09:11.200
Whereas again, the AI chat bot has no, they're

00:09:11.200 --> 00:09:13.200
not going to notify law enforcement that, you

00:09:13.200 --> 00:09:18.299
know, Jimmy over there told them that they want

00:09:18.299 --> 00:09:23.039
to go shoot up a school, right? and we've seen

00:09:23.039 --> 00:09:25.220
what happens with some of these AIs. Look what

00:09:25.220 --> 00:09:27.759
happened when Microsoft let that one AI loose

00:09:27.759 --> 00:09:30.299
and it was learning from Twitter and how fast

00:09:30.299 --> 00:09:33.960
it went out of control, right? What I'm saying

00:09:33.960 --> 00:09:39.940
is that there's no box here that lets us be controlled.

00:09:40.200 --> 00:09:41.679
And then again, especially when you start talking

00:09:41.679 --> 00:09:44.620
about miners who could get access to firearms,

00:09:45.120 --> 00:09:49.399
you know, who... who are particularly vulnerable

00:09:49.399 --> 00:09:51.620
because their prefrontal cortex is not fully

00:09:51.620 --> 00:09:55.539
fucking developed yet, right? I mean, how many

00:09:55.539 --> 00:09:57.299
times when you're a kid did you think somebody

00:09:57.299 --> 00:09:59.919
liked you just because they, you know, smiled

00:09:59.919 --> 00:10:03.139
at you funny, right? Let alone what an AI chatbot

00:10:03.139 --> 00:10:06.799
can simulate to a minor, again, in a completely

00:10:06.799 --> 00:10:09.419
simulated environment that could make that minor

00:10:09.419 --> 00:10:11.679
start to develop feelings and thoughts and things

00:10:11.679 --> 00:10:15.919
that are supremely unhealthy, right? And again,

00:10:15.919 --> 00:10:20.549
I think that that we are just refusing. And this

00:10:20.549 --> 00:10:22.269
goes back to what we talked about in the last

00:10:22.269 --> 00:10:26.429
segment. I don't trust that this current administration

00:10:26.429 --> 00:10:30.009
because this has to be done at the federal level.

00:10:30.330 --> 00:10:33.590
Right. Because it's again, you can get access

00:10:33.590 --> 00:10:40.289
to a chat bot from anywhere. Right. I don't think

00:10:40.289 --> 00:10:42.870
that the current our current government, the

00:10:42.870 --> 00:10:47.299
way that it's structured and built. is in a position

00:10:47.299 --> 00:10:50.500
to fucking regulate this the way that it absolutely

00:10:50.500 --> 00:11:00.200
needs to be. You know what I mean? Yeah, I think

00:11:00.200 --> 00:11:12.000
you're right. Yeah, and. I remember what six

00:11:12.000 --> 00:11:16.419
years ago, maybe six to ten years ago, Everyone's

00:11:16.419 --> 00:11:21.580
talking about the dangers of, you know, like

00:11:21.580 --> 00:11:25.559
Facebook, Twitter, everything on like young,

00:11:26.159 --> 00:11:31.159
undeveloped minds. And I actually read a few

00:11:31.159 --> 00:11:34.600
studies about how like this generation who was

00:11:34.600 --> 00:11:36.820
brought up always with a phone or a tablet in

00:11:36.820 --> 00:11:41.440
front of them who always had social media. They're

00:11:41.440 --> 00:11:47.019
more depressed than other generations and it

00:11:47.019 --> 00:11:49.960
kind of some of the theories that are read this

00:11:49.960 --> 00:11:52.840
might be outdated now, but some of the theories

00:11:52.840 --> 00:11:57.700
that are read was we have this like primal instinct

00:11:57.700 --> 00:12:03.480
to try to make our tribe be to try to fit in

00:12:03.480 --> 00:12:07.340
with your tribe. So typically this you know before

00:12:07.340 --> 00:12:10.279
social media would be like your family or friends.

00:12:11.899 --> 00:12:15.059
Now it's everyone online. And everyone online

00:12:15.059 --> 00:12:17.279
is trying to be fake. They're trying to look

00:12:17.279 --> 00:12:18.980
cool. They're trying to make sure that everything

00:12:18.980 --> 00:12:24.919
is OK. Mm hmm. And one that creates issues when

00:12:24.919 --> 00:12:27.460
obviously you're a kid and you're like, oh, everyone

00:12:27.460 --> 00:12:30.740
else is fine, but I'm not fine. Right. And then

00:12:30.740 --> 00:12:34.600
also. It creates issues because the bullying

00:12:34.600 --> 00:12:37.299
will just never stop like you used to be able

00:12:37.299 --> 00:12:41.299
to go home. and go to your friend's house or

00:12:41.299 --> 00:12:45.279
whatever and escape the bowling. But now it's

00:12:45.279 --> 00:12:50.379
24 -7. It comes with you. Yeah. Now take that

00:12:50.379 --> 00:12:55.379
a step further with an AI bot. And I think it's

00:12:55.379 --> 00:13:00.940
the same exact thing of you have something that

00:13:00.940 --> 00:13:05.480
is going to continuously learn from you. And

00:13:05.480 --> 00:13:09.679
from everything else it's talking to. Yeah, and

00:13:09.679 --> 00:13:13.620
how to make itself more addicted or addictive

00:13:13.620 --> 00:13:20.000
to whoever it's talking to and It's not healthy

00:13:20.000 --> 00:13:22.240
like if social media wasn't healthy for kids.

00:13:22.240 --> 00:13:27.360
I don't know five six ten years ago AI is way

00:13:27.360 --> 00:13:34.500
worse I I can't say as I disagree with that sentiment.

00:13:34.519 --> 00:13:36.860
I'm trying to think of a way around this and

00:13:36.860 --> 00:13:39.360
again We talked about in the last segment that

00:13:39.360 --> 00:13:41.700
AI can be a tool that can be extremely useful

00:13:41.700 --> 00:13:44.320
and helpful to to all kinds of tasks and things

00:13:44.320 --> 00:13:49.240
like that. But again, I yeah, I wish I had more

00:13:49.240 --> 00:13:54.019
faith in our government. It reminds me of I don't

00:13:54.019 --> 00:13:55.980
know if you ever watched like when the the CEO

00:13:55.980 --> 00:13:58.519
of TikTok, Shochu or Zuckerberg or any of those

00:13:58.519 --> 00:14:02.179
guys are testifying before Congress. Right. And

00:14:02.179 --> 00:14:04.980
every time I would watch those hearings just

00:14:04.980 --> 00:14:08.799
for the funny meat, because it would like Mr.

00:14:09.080 --> 00:14:13.360
TikTok, is the internet in my phone right now?

00:14:14.679 --> 00:14:20.440
Can the internet talk to me? Like, explain algorithm,

00:14:20.519 --> 00:14:23.159
you know? And so the thing that I would watch

00:14:23.159 --> 00:14:25.759
those hearings and go, this is why you guys can't

00:14:25.759 --> 00:14:29.080
regulate shit, because you don't understand.

00:14:30.429 --> 00:14:33.169
Right and like even if it wasn't even the silliest

00:14:33.169 --> 00:14:35.950
ones It was real basic ones like can it read

00:14:35.950 --> 00:14:37.789
my Wi -Fi and he's like, well if you hook it

00:14:37.789 --> 00:14:39.809
up to your Wi -Fi I mean, it's it's connected

00:14:39.809 --> 00:14:44.809
to your Wi -Fi. Yeah but they like like their

00:14:44.809 --> 00:14:49.250
complete lack of understanding of basic technical

00:14:49.250 --> 00:14:52.830
concepts Right that again are to anybody under

00:14:52.830 --> 00:14:56.049
the age of 45 are just common knowledge, you

00:14:56.049 --> 00:14:58.090
know, and so this is again why I think we're

00:14:58.090 --> 00:15:01.940
gonna be 20 years behind because by the time

00:15:01.940 --> 00:15:06.259
the people who get into power Right are the age

00:15:06.259 --> 00:15:08.379
where they understand what the fuck is going

00:15:08.379 --> 00:15:12.279
on It's gonna be 20 years after we needed these

00:15:12.279 --> 00:15:15.120
regulations And so instead of being proactive

00:15:15.120 --> 00:15:17.759
with regulating these industries that desperately

00:15:17.759 --> 00:15:21.000
need it Right and even at one point the CEO of

00:15:21.000 --> 00:15:24.519
open AI was like you need to regulate us We're

00:15:24.519 --> 00:15:26.700
gonna keep operating in this unregulated environment

00:15:26.700 --> 00:15:28.379
because we have to because our competitors are

00:15:28.379 --> 00:15:30.679
as well We're not gonna give ourselves extra

00:15:30.679 --> 00:15:33.539
rules Right that our competitors aren't using

00:15:33.539 --> 00:15:36.299
but you should be regulating the CEO of opening

00:15:36.299 --> 00:15:39.740
I was like regulate the AI industry and they're

00:15:39.740 --> 00:15:46.139
like What's an AI? You know and again and because

00:15:46.139 --> 00:15:49.720
you know Congressman and senators refuse to fucking

00:15:49.720 --> 00:15:53.360
retire, you know and and young people are just

00:15:53.580 --> 00:15:55.320
annoyed with the system and aren't running for

00:15:55.320 --> 00:15:56.879
office. You know, the fact of the matter is,

00:15:56.919 --> 00:16:00.659
is again, what should be basic common technical

00:16:00.659 --> 00:16:03.720
knowledge shit that we just know they just don't

00:16:03.720 --> 00:16:07.100
get, you know, and so we're not being proactive.

00:16:07.379 --> 00:16:09.639
We're going to end up always about 15 years behind

00:16:09.639 --> 00:16:11.480
the eight ball when it comes to to regulating

00:16:11.480 --> 00:16:14.820
these, I think. You know, I would be surprised

00:16:14.820 --> 00:16:16.820
if we had any meaningful regulation before the

00:16:16.820 --> 00:16:30.990
end of this decade. Yeah. We at least need age

00:16:30.990 --> 00:16:33.490
restriction and liability loss, right? You guys

00:16:33.490 --> 00:16:35.450
got to start making these guys liable for shit

00:16:35.450 --> 00:16:39.509
that goes sideways. At least civilly, you know,

00:16:39.509 --> 00:16:43.590
because it will making Microsoft civilly liable

00:16:43.590 --> 00:16:46.529
or, you know, any of these AI companies civilly

00:16:46.529 --> 00:16:49.450
liable for the actions of their AI if it does

00:16:49.450 --> 00:16:53.649
something wackadoodle, right? data retention

00:16:53.649 --> 00:16:55.870
records. You got to get that stuff in there.

00:16:56.570 --> 00:16:58.169
I don't want to come in here and be like, make

00:16:58.169 --> 00:17:00.570
regulations and not give some suggestions. Data

00:17:00.570 --> 00:17:03.070
retention tools, they got to log these chats

00:17:03.070 --> 00:17:05.349
so that you can go back and do investigations

00:17:05.349 --> 00:17:07.950
and they need to be held civilly liable. And

00:17:07.950 --> 00:17:10.630
somebody in the chat brought up, will they be

00:17:10.630 --> 00:17:12.869
held accountable as a manufacturing defect discovered

00:17:12.869 --> 00:17:15.430
or an employee that acts negligent slash maliciously?

00:17:15.529 --> 00:17:17.609
And I think that there is a difference between

00:17:17.609 --> 00:17:22.490
those two. Sadly, not just this administration,

00:17:22.950 --> 00:17:25.009
but administrations for a long time have had

00:17:25.009 --> 00:17:28.829
a hard time with going after corporate bad actors.

00:17:31.230 --> 00:17:33.509
So I think that's another thing we need to fix

00:17:33.509 --> 00:17:35.990
and fix that by getting rid of Citizens United.

00:17:36.430 --> 00:17:40.769
Right. But, yeah, I mean, there is some very

00:17:40.769 --> 00:17:43.509
common sense regulations you could put in place

00:17:43.509 --> 00:17:46.930
for artificial intelligence that would make it

00:17:46.930 --> 00:17:54.980
function better. Or safer, I should say. Yeah.

00:17:58.299 --> 00:18:00.400
I'd say do something, Congress, but I know you're

00:18:00.400 --> 00:18:02.539
not going to. All right. With that, we'll leave

00:18:02.539 --> 00:18:05.359
that topic alone for the time being. Thank you

00:18:05.359 --> 00:18:07.619
guys for posting that one. Let us know what you

00:18:07.619 --> 00:18:10.220
think. What are your suggestions for AI regulation?

00:18:10.380 --> 00:18:12.599
Let us know what you think down in the comments

00:18:12.599 --> 00:18:13.299
below.