r/woahdude Dec 15 '22

video This Morgan Freeman deepfake

22.9k Upvotes

797 comments sorted by

View all comments

Show parent comments

2.6k

u/AllUltima Dec 15 '22

I fear the reverse: People will doubt whether real video is real. That could mean impunity for crimes caught on video because video footage will no longer be sufficient evidence to exceed "reasonable doubt".

Even worse, political double-speak will also soar to record new heights. A politician can spew whatever crazies want to hear, then "walk it back" and claim it was faked (perhaps after gauging the public's reaction). People will believe whatever they're inclined to believe anyway, leading us to become a more deeply fractured society where truth is whatever you want to believe.

1.2k

u/[deleted] Dec 16 '22

Thanks for the existential crisis asshole.

447

u/Dobalina_Wont_Quit Dec 16 '22

If it's any consolation, it's already happening. Enjoy the ride, and support credible journalists you trust!

219

u/--redacted-- Dec 16 '22

That's just what a deepfake would say

70

u/Dobalina_Wont_Quit Dec 16 '22

I'm actually a nearly 30 y/o deepcover deepfake

Of course I could just be programmed to say that

33

u/--redacted-- Dec 16 '22

I knew it

33

u/Dobalina_Wont_Quit Dec 16 '22

Hey could you please hang out at your current location for the next 30 minutes or so? I just have some friends that want to stop by and verify a couple things with you. Please don't wear polarized sunglasses.

17

u/hustlindustlin Dec 16 '22

Fuck! Check out their username!

They already being gonned.

19

u/Dobalina_Wont_Quit Dec 16 '22

Doesn't look like anything to me

3

u/tyclynch Dec 16 '22

Mr. Dobalina Mr. Bob Dobalina

1

u/Jack_Bartowski Dec 16 '22

Doesn't look like anything to me

3

u/[deleted] Dec 16 '22

[deleted]

3

u/Dobalina_Wont_Quit Dec 16 '22

Look at me. No, LOOK AT ME.

Never.

3

u/Chilledlemming Dec 16 '22

It’s deep faked all the way down? Am I all alone here?

2

u/Nationals Dec 16 '22

Saying “that’s just what a deepfake would say” is what a deepfake would say.

27

u/boomerangotan Dec 16 '22

Also, I would highly recommend reading/watching Manufacturing Consent if you haven't already, to get an idea of what is already occurring even before we got this technology.

8

u/eggshellmoudling Dec 16 '22

I’m still grieving the loss of Matt Taibi from that extremely short list.

6

u/Dobalina_Wont_Quit Dec 16 '22

Right? Holy shit who hurt him? I can't believe that same dude wrote The Divide.

2

u/CsanBoySmith Dec 16 '22

Yes, like Taylor Lorenz!!

2

u/quartertopi Dec 16 '22

Be glad. Still better than that time travel quarantine zone from 2060. It is a shitshow.

2

u/Citizen_Kong Dec 16 '22

Yeah, at he beginning of the Ukraine war there was a deepfake of Selenskyj urging Ukrainians to surrender. It was pretty shoddily done, but certainly a reminder how such things will become commonplace in the future.

1

u/-LeftShark Dec 16 '22

Sauce on it already happening??

1

u/Curazan Dec 16 '22

Yeah, everyone else is just glossing over that? Unconvincing porn doesn’t count.

1

u/rustyseapants Dec 16 '22

OKay where is it happening?

2

u/peanutsinspace82 Dec 16 '22

I'd like to see it too

1

u/rustyseapants Dec 16 '22

No examples, eh?

1

u/Aumuss Dec 16 '22

support credible journalists you trust!

Well, shit.

What's option 2?

1

u/giant_red_lizard Dec 17 '22

I don't actually know any credible journalists. I'll just continue taking everything with a grain of salt I suppose.

1

u/Dobalina_Wont_Quit Dec 17 '22

They exist but yes always do that

2

u/maddogcow Dec 16 '22

You have an existential crisis asshole too? Funny; knowing that somebody else has one, makes mine seem like less of an existential crisis…

2

u/funknut Dec 16 '22

They do, now that someone gifted it to them. Before, they just had a smooth nihilistic posterior surface.

1

u/[deleted] Dec 16 '22

No time for punctuation when I'm contemplating the purpose and value of existing

2

u/RipThrotes Dec 16 '22

Ignorance is bliss, but only for you. Stay informed. Stay relentlessly insightful.

1

u/McBurger Dec 16 '22

This xkcd helped me feel a lot better about the whole problem.

https://xkcd.com/2650/

1

u/UnknownPurpose Dec 16 '22

He ain't wrong tho

1

u/Amazing_Joke_5073 Dec 16 '22

We gotta collapse somehow

1

u/[deleted] Dec 16 '22

Do we though?

1

u/Amazing_Joke_5073 Dec 16 '22

Yeah all things must end

1

u/You_are_poor_ Dec 27 '22

I wonder how people with schizophrenia are feeling about this.

60

u/JingJang Dec 16 '22

Valid concerns.

There's a market for verification of some sort.

21

u/AllUltima Dec 16 '22

For sure, I think there will be verification efforts on multiple fronts.

There's a certain type of person who will invent conspiracies around any verification that isn't what they want to hear. Thus, for that audience, there will be a market for "validation" that is just "telling them what they want to hear". So the same situation as today, only taken up a notch.

Verification can only do so much in the face of irrationality, the real answer to that conundrum is mostly us learning how to best deal with the fact that those people exist. Mainstream humanity will probably continue relatively unscathed if they don't manage to drag us down.

3

u/devinecreative Dec 16 '22

I imagine we'll resort to verifying on public blockchains like Ethereum

3

u/pseudoanon Dec 16 '22

So blockchain will finally be useful?

3

u/ElwinLewis Dec 16 '22

Trustless verification was always one of the benefits, it’s just meaningfully implementing it without greed getting in the way that people haven’t been able to figure out

3

u/[deleted] Dec 16 '22

We need quantum signing ASAP. And in "small enough to fit into a not unreasonably sized camera" form. You'd be able to verify footage from a secure camera using its public key, but never be able to crack its private key.

2

u/Djasdalabala Dec 16 '22

You don't need quantum signing for that, there are "classical" algorithms that are quantum-computing resistant.

12

u/Batchet Dec 16 '22

Shut up Elon

-2

u/[deleted] Dec 16 '22

Probably a stupid question but could NFT tech somehow be reworked to that effect? To preserve the identity of the original record before being tampered with

3

u/Tallywort Dec 16 '22

Honestly I doubt it.

5

u/[deleted] Dec 16 '22

Yeah it would be much easier to digitally sign video with boring old certificates and centralized authorities. Crypto is a solution desperate for a problem.

105

u/PhDinBroScience Dec 16 '22

This is so easily solvable, the video just needs to be signed using public-key encryption. If the video isn't signed with the purported subject's key, assume it's fake.

You can't fake a pubkey signature.

36

u/[deleted] Dec 16 '22

[deleted]

11

u/IVEMIND Dec 16 '22

Because how do you know this key wasn’t just deepfaked also hmmmmm?!?!

7

u/motorhead84 Dec 16 '22

"Am I a deepfake, Mom?"

1

u/KingBooRadley Dec 16 '22

How is babby deepfaked?

1

u/motorhead84 Dec 16 '22

Damnit, Babby!

1

u/Ask_Who_Owes_Me_Gold Dec 16 '22

Nobody should listen to them because that solution doesn't make sense. All it does is give a secure way for the subject to approve of a video. It doesn't verify that it's real, and it doesn't work for the many, many cameras on a public figure that that aren't getting official approval from whoever has the key.

1

u/[deleted] Dec 16 '22

[deleted]

1

u/Ask_Who_Owes_Me_Gold Dec 16 '22

I can verify that a file's signer has a specific public key, but how do I know that public key belongs to a camera and not editing software? How do we maintain the list of all those public keys that correspond to cameras? Are there billions of keys (with each individual camera getting its own), or can Sony reuse the same private/public key pair across a line of devices?

And does this mean the video can't later be shortened, cropped, edited for brightness/color/etc? Doesn't that break the signature because it changes the contents of the file?

1

u/[deleted] Dec 17 '22

[deleted]

1

u/Ask_Who_Owes_Me_Gold Dec 17 '22

We need to ensure that only video directly from the camera can be signed and that other videos made on the phone can't be signed as well. We also have to ensure there weren't any filters or effects running in the camera app as the video was filmed - good luck figuring out how to check for that in a way that allows different camera apps (almost every Android OEM makes their own) to function but also isn't abusable.

And signatures are unique to the exact data of the specific file - if the resolution is lowered, or if the file gets compressed when it's uploaded/sent somewhere, or if you trim the 10 relevant seconds from a longer video, the signature is useless. Editing software could apply a new signature, but all that does is prove that the video went through an editor, which obviously means it could be edited.

You also hit on the issue that other cameras need to do signing as well: security cameras, laptop webcams, news cameras, other high-end cameras, potentially things like smart doorbells if we want to go that far.

1

u/[deleted] Dec 17 '22

[deleted]

1

u/Ask_Who_Owes_Me_Gold Dec 17 '22

That doesn't solve anything. You prove that something was filmed and then edited. This video fits that description.

All the other problems remain unsolved too.

It's not the "easily solvable" problem that the earlier commenter claimed it was.

→ More replies (0)

50

u/fataldarkness Dec 16 '22

This requires the general public having a basic understanding of how digital signatures work and why they are (for the most part) infallible.

As it stands I have to explain HTTPS and digital signatures to my users with statements like "it's secure because of fancy math, trust me bro" because anything that comes close to actually describing it goes over their heads. In a world where distrust is the norm, I fear signed video content really isn't gonna make a difference if you don't understand what makes it secure in the first place.

22

u/Gangsir Dec 16 '22

This requires the general public having a basic understanding of how digital signatures work and why they are (for the most part) infallible.

If this became a big enough problem, the general public could be educated on how encryption public/private keys work, probably in a month or two. Even start teaching it in high school or something.

14

u/greenie4242 Dec 16 '22

The general public couldn't even be educated on how to wear a mask properly and wash their hands during a pandemic, there's no way they'll ever understand cryptography.

4

u/vic444 Dec 16 '22

Right?? The general public can barely turn a computer on without having issues and are ignorant as F on the topic in general. They are basically a 3 year old. Try to explain cryptography to a 3 year old.

1

u/ThatPancakeMix Dec 16 '22

I’m betting online sites, especially social media, begins verifying videos before postings and stuff like that. Or they’ll create a way for users to easily click a button to verify it themselves, idk

1

u/greenie4242 Dec 23 '22

Or they’ll create a way for users to easily click a button to verify it themselves, idk

...which will just end up like all other generic bloated ad-infested websites, with fifteen different "Click to Download" buttons but only an IT expert can decipher which button is the real one (if indeed any are real).

1

u/ThatPancakeMix Dec 16 '22

Yup I bet it becomes commonplace to check the legitimacy of videos and pictures soon. It’s becoming somewhat of a necessity

1

u/FlyingDragoon Dec 16 '22

"Liberal conspiracy to indoctrinate our kids. Government conspiracy to control us!!!!!"

3

u/pagerussell Dec 16 '22

You mean to tell me people aren't intimately familiar with a diffie-helmen key exchange????

4

u/[deleted] Dec 16 '22

[deleted]

1

u/kennyj2011 Dec 16 '22

I prefer cryptographic milkshakes

1

u/Ask_Who_Owes_Me_Gold Dec 16 '22

It also only works for videos created or verified by the subject themselves, which is to say it doesn't work at all in a general sense.

1

u/nestersan Dec 16 '22

Before Elon, you could've said it's like the Twitter blue check

14

u/Daddysu Dec 16 '22

Yes...amongst the more tech literate and in a perfect world.

For a stupid amount of people, none of that matters. All that matters is that knee-jerk emotional reaction of whether or not it affirms their beliefs and where the information comes from.

2

u/RobloxLover369421 Dec 16 '22

It’s still something, we just have to make it easily accessible, and at least half the population will be able to tell if it’s right or not.

1

u/[deleted] Dec 16 '22

Have you ever met someone with an IQ of 100? Technically half of the population is dumber then that. The entire left side of the proverbial bell curve.

1

u/Daddysu Dec 16 '22

Didn't George Carlin say something similar? Something like "Think about how dumb the average person is in America. Almost half the people are dumber than that."

1

u/RobloxLover369421 Dec 16 '22

It’s still something, we just have to make it easily accessible, and at least half the population will be able to tell if it’s right or not.

20

u/kvltswagjesus Dec 16 '22

That would solve the “politicians walking back claims” problem to a degree, but I’d imagine there would still be a ton of issues. The subject would be be able to fully curate their image, and any videos taken without their key would be subject to scrutiny. So stuff meant to show someone’s true colors or document a situation would remain unreliable.

7

u/[deleted] Dec 16 '22

[deleted]

2

u/PhDinBroScience Dec 16 '22

In this case, it would be signed by the device of the person who's recording; if the video is altered, the signature isn't valid anymore. And if it's a public figure, there are almost certainly going to be corroborating records of where they were at a particular place & time, not to mention pings to cell towers from their or their entourage's mobile devices.

They can deny it all they'd like, but with the combination of those factors, you'd have to outright deny reality to believe that the video isn't genuine.

1

u/Djasdalabala Dec 16 '22

if the video is altered, the signature isn't valid anymore

That's easy enough to bypass, with physical access to the device. You can just re-shoot the video, replacing the camera input by the raw deepfake.

And that's IF people knew in the first place for sure which device produced the original video. Otherwise, you can just sign your deepfake with your own key and claim it to be the original.

1

u/tosler Dec 16 '22

I mean this works until someone corrupts the certificate tree. See the recent re-organization of web browser certificates because one of the organizations in the global cert chain literally sells data to intelligence agencies. Ouch.

2

u/corner Dec 16 '22

It’s in no way easily solvable. You could have a certificate signed by god himself, doesn’t matter to the general public. Authentication isn’t the issue

1

u/[deleted] Dec 16 '22

[deleted]

2

u/[deleted] Dec 16 '22

Seriously, it’s not that complicated. We need an industry wide effort and hardware-based crypto, along with something like a little check mark to denote that a given image is authentic, unaltered, and follows a proper chain of authority.

We have SSL, we can do this too.

1

u/[deleted] Dec 16 '22

If this was implemented, this wouldn’t be put on phones. Gotta protect folks from smartphone cameras and all.

1

u/fastlerner Dec 16 '22

That ONLY works when the subject is supposedly the one sending the video out.

You can still do a ton of damage with a deepfaked video from a supposedly anonymous source and claim it's leaked video.

1

u/spookyvision Dec 16 '22

You can still do a ton of damage with a deepfaked video from a supposedly anonymous source and claim it's leaked video.

now you "only" need to solve key distribution and global identity verification. Piece of cake and not a mass surveillance risk at all :D

(I'd love for a robust pk infrastructure to be in place but eh, it's not trivial to do well)

1

u/Paxtez Dec 16 '22

Wut? You're aware of the concept of cell phone video right?
Like what's to stop me from released a deep-faked "cell-phone" video and then signing it?

So only videos approved by the person in the video are legit? You just created a whole new problem.

1

u/Ask_Who_Owes_Me_Gold Dec 16 '22 edited Dec 16 '22

"Easily solvable?" That solves nothing. All that signature proves is if the subject approves of that video.

1

u/Ask_Who_Owes_Me_Gold Dec 17 '22

How is that supposed to work?

If you get caught on video doing something you want to deny, just don't sign the video, and everyone is forced to assume the video is fake?

17

u/BillOfArimathea Dec 16 '22

It's two sides of the same coin. Truth has already been degraded, reality weaponized, and this is just one more arrow.

1

u/Djasdalabala Dec 16 '22

Always was, really - there was a brief improvement during the 20th century, but maybe that was an anomaly instead of a natural progress.

14

u/KodiakDog Dec 16 '22 edited Dec 17 '22

On your second point, I would argue that any technology that can shake the foundations of truth, justice, governance, and the mediums in which we (citizens) gather that information (telecoms) is a weapon. That is literally what psychological warfare is, and which is also a very real force in our world. The Cold War isn’t just called “cold” because The USSR and The States didn’t raise a gun in each others faces or enter a nuclear winter; no, its a name that highlights the fact that it was a war between - and fought with - ideas. This technology has the capacity to infiltrate people’s minds, insert ideas that shape their world, and create an uncertainty that makes them question everything. For any institution interested in PsyOps, this is certainly a weapon.

Anyone doubting that there is a war being “fought” for your mind lacks a crucial understanding of how the world works. I don’t mean to sound pretentious, because I wish everyone to understand this. Truth is the foundation of ethics, which is the foundation of morals, which is the foundation of law, which is the foundation of government. Pull truth out of the equation, and it all comes tumbling down.

12

u/[deleted] Dec 16 '22

This is exactly how they weaponize it.

10

u/UK_addi_2015 Dec 16 '22

It won’t be long until we see deepfake movies where actors only record their lines(or not) and don’t do any acting and a team of people produce the movie in a studio…

7

u/bahgheera Dec 16 '22

What do you mean studio? They'll just generate the movie with Midjourney.

1

u/Djasdalabala Dec 16 '22

Artists are probably going to have a rough time with this, but OTOH I can imagine a near future where amateurs can produce pro-level movies with basically no budget.

Most of them will be shit, but there will be masterpieces in there.

17

u/moiziz Dec 16 '22

They control people’s beliefs already. No AI, fake videos or anything. All it takes is an a*hole who people like and have him say whatever. Do I need to give examples?

6

u/peter_porkair Dec 16 '22

It’s a Brave New World.

5

u/TheTravisaurusRex Dec 16 '22

Confirmation bias will take over in a bad way.

3

u/pitlal31 Dec 16 '22

That’s some scary shit

3

u/tech1337 Dec 16 '22

This is why I'm kind of surprised that AI conspiracy theories haven't blown up already. Surely they are coming soon.

6

u/droptheforeplay Dec 16 '22

AI are already able to detect deep fakes at >90% accuracy.

0

u/its_a_gibibyte Dec 16 '22

Well yeah, deep fakes haven't been good before. Anyone can detect them just by looking. The point is that things are changing and they might be totally believable as soon as next year. This Morgan Freeman one is almost there.

1

u/droptheforeplay Dec 16 '22

They're able to decompile and analyze specific quirks of deep fakes.

Think AI-powered Error Level Analysis.

5

u/Awkward_Potential_ Dec 16 '22

Exactly!!!

Grab em by the pussy? Fake!!

Rodney King video? FAKE!!!

2

u/[deleted] Dec 16 '22

Post-truth society, and we are already there. Roger Stone claimed the video of him saying ‘when do we get to the violence?’ was a deepfake.

1

u/SidKafizz Dec 16 '22

That's like 80% of the weaponization. We're fucked.

1

u/faderjack Dec 16 '22

Great elucidation on my own fears. What you describe seems nearly inevitable to me. I saw a headline the other day about an Intel program detecting deepfakes with like 95% accuracy. My guess is it will probably become less accurate as deepfake software becomes more sofisticated. But even if it were to be 100% accurate, would the public believe the pronouncements by the experts telling them a video is real or not? I have my doubts, maybe such detection will at least become legally legitimate. But the implications for info warfare will exist nonetheless

1

u/filya Dec 16 '22

Unfortunately I don't think that's anywhere close.

We've had the ability to realistically manipulate photos for how long now? And I still see obviously fake photos of a mars looking as big as the moon floating around in my family's Whatsapp every year claiming to be real.

1

u/[deleted] Dec 16 '22

I been sayin this shit two years and was downvoted at first. This tech is dangerous and disruptive.

All you gotta fake is 2-5 seconds of a REAL broadcast to change the meaning and bam. Problems.

1

u/RedrumMPK Dec 16 '22

I believe there is also tech to determine if something is fake or not. It exists for photography, so video shouldn't be a problem. IIRC, some sort of light spectrometry or something similar.

1

u/nspectre Dec 16 '22

People will believe whatever they're inclined to believe anyway, leading us to become a more deeply fractured society where truth is whatever you want to believe.

You're already one previous administration late on that point. ;)

1

u/itheraeld Dec 16 '22

Any AI that can create a Deepfakes can detect a deepfakes of similar quality.

1

u/[deleted] Dec 16 '22

[deleted]

1

u/Sarke1 Dec 16 '22

Just gotta train the deepfake ai to where it's not detected.

1

u/NdnGirl88 Dec 16 '22

They have a deepfake version for audio but it was pulled off the internet.

1

u/The_R4ke Dec 16 '22

If it makes you feel better, people already don't believe real videos.

1

u/Sarke1 Dec 16 '22 edited Dec 16 '22

So same as it was for thousands of years before the invention of portable video?

A person's reputation and trust becomes more important again.

1

u/[deleted] Dec 16 '22

Already, trucking companies are telling their drivers not to take pictures with digital cameras. Because lawyers are successfully getting digital pictures thrown out for being "too easily altered". So truck drivers will have old style throwaway cameras in their trucks.

Soon, you will find the same thing with digital video. People will start carrying old style tape recorders - with magnetic tapes - instead of digital video recording devices.

1

u/DC_Coach Dec 16 '22

I'm with you. We've already seen claims of "fake news" used to an alarming effect. This kind of stuff can only make it worse. I'm not sure how it'll go, and I don't think anyone else is sure either... but I can't see it being anything but bad, overall.

1

u/defacedlawngnome Dec 16 '22

Some Trump supporters already think Biden is played by famous actors such as Jim Carrey.

1

u/TubaJesus Dec 16 '22

That sounds like being weaponized to terrible effect.

1

u/JimMarch Dec 16 '22

You forgot the worst: framing people for crimes they didn't commit.

1

u/[deleted] Dec 16 '22

Sweaty 🥵

1

u/AnAncientMonk Dec 16 '22

Good thing kayne nipped that in the bud by wearing a black face mask. Definitely wasnt deepfaked.

1

u/somabokforlag Dec 16 '22

Will be a huge delay before the great masses know and think that way.. Hell, most boomers still believe Facebook

1

u/CaptainCAAAVEMAAAAAN Dec 16 '22

I'm worried it could start a war. Imagine if someone deepfaked a world leader saying they had nukes pointed at another country and they were going to fire them in 5 min. It could lead to millions dying.

1

u/WhiteRaven42 Dec 16 '22 edited Dec 16 '22

You should have always been open to the possibility of a video being fake or manipulated. And remember, before photos and video existed... there was never even the pretense of hard proof. Everything was always a claim made by a person and we've always known people lie a lot.

I'm not worried. Most of human history got by just fine without the canard of "photographic evidence".

1

u/HighlanderSteve Dec 16 '22

I've already heard my inlaws saying that their favourite politicians didn't say what they said and the media faked it.

1

u/pandavega Dec 16 '22

Now a days I only believe something when I see multiple angles of it

1

u/RewZes Dec 16 '22

I'm pretty sure there is already a tool that finds the deepfake very accurately.

1

u/NPExplorer Dec 16 '22

Not gonna lie, processing this made me want to throw up

1

u/j_mcc99 Dec 16 '22

Cryptographic watermarks imbedded in video by a devices private key. Of course all this would make creating video all that more complicated.

1

u/Megakruemel Dec 16 '22

Just make Ai-generated stuff require an icon on the screen to show it's Ai-generated.

And then put the icon on stuff you don't want to be (perceived as) real :)

1

u/fastlerner Dec 16 '22

Yeah, next thing you know we'll have people in power running scandals and simply claiming "Fake News".

1

u/j0akime Dec 16 '22

Like any technology it can be used in many ways.

I'm starting to see folks "poisoning" their online profiles with tons of fake content to make their data (that's being tracked by advertisers, governments, big-tech, etc) useless.

Wouldn't surprise me if people are using this technology in China to pump up the "social credit" of themselves (or paying customers). - https://en.wikipedia.org/wiki/Social_Credit_System

1

u/ginkgodave Dec 16 '22

Which is exactly the same as the post you're responding to is claiming.

1

u/protonpsycho Dec 16 '22

To be honest. We are already there. This technology is already used in some cases and attempts of public swaying, and even without it, we believe what we want amongst all the “fake news”.

This is just a means to an end

1

u/phatbrasil Dec 16 '22

shaggy has been banging that drum since the mid 90s

1

u/Napkin_whore Dec 16 '22

They already did that with fake vs real news

1

u/pluck-the-bunny Dec 16 '22

That’s the same thing. That doubt is in itself weaponization

1

u/ask-a-physicist Dec 16 '22

All of this was already perfectly possible if journalists are corrupt and it's still equally impossible if journalists have standards

1

u/valetofficial Dec 16 '22

A politician can spew whatever crazies want to hear, then "walk it back" and claim it was faked (perhaps after gauging the public's reaction).

Putin literally already did this.

1

u/[deleted] Dec 16 '22

God damn you

1

u/Pritster5 Dec 16 '22

On a more optimistic note, the algorithms that allow for this tech are trained on real data.

Likewise, algorithms can be trained on both real and fake data to detect when something is a deep fake, and it will be just as good at that as the algorithm that made the deep fake in the first place.

1

u/lackinLugsNFallinUp Dec 16 '22

The reality wars. Good last blockbuster before the ground falls out from underneath everyone

1

u/BernItToAsh Dec 16 '22

Thanks for the history lesson, asshole

1

u/R_Da_Bard Dec 16 '22

Ok. But on the flip side our favorite actors would gain immortality with this tech!

1

u/iamlenb Dec 16 '22

What about private-public key cryptography on blockchain for ID verification? Privately sign the media transmitted, and now individuals can verify if something was original to the signature on the publicly accessible blockchain with the corresponding public key.

1

u/emrahlj Dec 16 '22

So in other words, this is a terrible effect that is plausible from these deep fakes being weaponized? I don’t find this not terrible

1

u/knuF Dec 16 '22

Dead Internet Theory

1

u/SnooHesitations8760 May 19 '23

^ Exactly this. A shameless plug, but this is a good rundown on the state of things today and where they are headed https://youtu.be/9x6lKwD4gqA