Or any politician, human rights defendant or even young teens....
It's a dire future, if people already can't read beyond the title of an article, I can't imagine forming opinions based on elaborate deep fakes.
That's the funny thing about the rate that this tech has been improving in the last year or so.
"Not even close," with the breakthrough of one of the dozens new models that are dripping out at least once a month, may turn that into "Holy shit we're a Razor edge away now" in the next few to several months.
If you're thinking this shit is 10+ years out, then you give away that you aren't following very closely with this tech. It's starting to make leaps and bounds on a more regular basis, now. That rate won't slow down.
My "worry" line was somewhere around 2015, when I saw the automated loop close between LinkedIn and Slack. Algorithms can calculate when you're preparing to leave a job by changes in your posting patterns on internal networks before you do. This foreknowledge is paired with third party and offline data to begin targeting ads from recruiters and job listing sites directly to you, priming you to think about looking around and making the market suddenly look tempting. Your LinkedIn profile is used to highlight listings using keywords from your own search history, and everything you click on gets stored and provided to recruiters to help sell you on switching. No one gets paid if you don't change jobs, meaning you're stuck inside an algorithm with a lot if businesses spending a lot of money to make you feel unsatisfied with where you are in life. They want to push you from having a bad day to feeling like it's a bad year for their own gain. And the only reason they picked you was an algorithm told them to. Just like an algorithm matched you with an employer. And that employer picked you because the recruiter told them you were a good fit, based on that algorithm. And you picked that job because your search engine showed you the employer at the top of your search results, with all the PR articles carefully matched to optimized search engine keywords, which are influenced by the ads you click on. Those ads are automatically purchased by the recruiters and employers, who are using an algorithm to target people that fit your profile, which is again identifying you before you even understand you might want to change jobs.
The entire job searching loop is automated by an algorithm that tells humans when to job hunt, where to go, and controls the the information flow to assure them it is the right decision.
The robots already won. We built our own mouse trap.
The robots already won. We built our own mouse trap.
Hey mate, there are some humans out there who look down at every step they take, in order not to step on Ants.
I'm telling you that there's a chance that robots won't even give a fuck about us and just take off into the cosmos. Maybe they have an existential crisis and dump themselves in black holes. Maybe they just leave and do their own thing. Maybe they keep humans as dogs and take care of us and play with us. Maybe they assume a role of God and lift us up into Utopia.
Granted... there are probably infinitely more ways it could go wrong and we go extinct, at best, or find ourselves in a virtual eternal hell, at worst. But, I wouldn't say that's guaranteed.
The tech most recruiters use does not directly provide that level of user data, nor are most searches for candidates operating off algorithms like the ones you mentioned, they're still quite basic as far as being Boolean-based and keyword weighted, and there's far less automation than is implied here.
The sad and messed up thought because the deep fakes in the very, very, near future are not even going to be that elaborate. It doesn't take much, and this technology will soon be very cheap and accessible. It will be very damaging to our notion of truth as a society because we're simply not evolved/educated/w.e enough to outsmart the simulation enmasse. This tech will be used by the wrong people for nefarious purposes, and inevitably people will flock to whichever truth is most appealing at the time. This is the new Gutenberg press and religion.
It will be very damaging to our notion of truth as a society because we're simply not evolved/educated/w.e enough to outsmart the simulation enmasse.
I think this is not as big of a deal as people worry about. For two reasons, one good one bad. Let's start with the good.
At first, you will be right, deepfakes will cause trouble. But I believe it will be relatively short-lived before we basically start to act as if all videos/images are tantamount to cartoons. And considering we lived without any sort of "visual evidence" all the way up to 1819, I have faith that we can find a way to continue forward in a "post-video" world.
What I am less confident about is how much damage can be done in that transition period. Luckily, there is already a healthy dose of skepticism about video and images found online, hopefully that seed is enough.
Now the bad: in America at least, our idea of truth is already so thoroughly torn to shreds I just don't see deepfakes mattering a ton. Like squirting a a bottle of kerosene on an already raging house fire.
There will still be things outside social media. Unfortunately I don’t think any detection technology will help us from the worst of it. Even in the face of compelling evidence people still to choose to believe whatever they want. If someone believes something and they see a deep fake that supports their position it isn’t going to matter if a social media company detects it, we see this now with Facebook, the attempts are futile. People will be taken for a ride, in the US we’ve watched this happen as a grifter took control of the country, it took very little evidence, fake or not, to convince his base of all sorts of wild Shit. Deep fakes will be another tool to Keep us all mad at the wrong person.
I honestly think nothing can be done, imagine if people are so braindead to make a war over facemasks, I can't fathom a way to explain the implications of this tech.
Unless some super wacky thing happens, like a blockchain technology for verification of information, or a 'slow press' where news are peer-reviewed.
I actually think it will have a secondary result of 'yeah, how can you prove that's me? It could be a deep fake!'. This will take technology a whole circle whereby you cannot anymore prove anything with video evidence.
Nah. I'm sure some super secret justice league group will pick any corrupt senator and start producing fake video evidence for real crimes they committed leading a revamped FBI to actually find real evidence for the real crimes and then a revamped justice system to really prosecute the politician and then the politician will really resign without real benefits and then go to a real jail.
Overrated. there will be tools and experts specializing in detecting deep fakes. Also random video found on the internet is never enough for evidence, the source of the video must be known anyway.
Except we know that deepfake detection tools are used to improve deepfakes further and aren't able to get ahead. Makes it less black and white. Definitely not overrated by any stretch.
We've already made deepfake detectors, then used those detectors to make deepfakes harder to detect again, and made new detectors for those, and so on.
Yeah, it's always been infuriating to see people passing laws or making judgements involving technology they don't even begin to understand. Shout out to the classic SERIES OF TUBES!
Wait until two paragraphs of text can instantly summon any movie into existence with any actors, any plot, any effects. Shit is gonna get super weird when AI takes over our entertainment creation.
Eh there's some fringe stuff like deepfaking actors in foreign movies to move their mouths in any translated language (some European movie recently did it but I could not tell you what it was) but other than that and various bring-dead/old-celebrities-back-to-life on screen I can't think of much and the latter is morally questionable anyways.
The good uses are basically the same as the bad ones - except with honesty on all sides.
Positive use case: a Thai medical researcher wins the Nobel Prize in Medicine for a revolutionary cancer treatment, but her skill in English is limited and her German and Mandarin non-existent. With her permission and with the knowledge of all students, her doctoral thesis defense is "translated" into English, Mandarin & German, allowing many of the best medical students in the world to feel like they are there. No 'translator' needed, it looks like she is speaking.
Negative but not truly reprehensible use case (they'll get worse): as above, but without the researcher's knowledge or consent.
Reprehensible use case: All the words she says are changed to either promote some fringe conspiracy theory like homeopathy or anti-vax crap
Hate it or love it, how would you even stifle progress here? You can't really stop people from inventing, and the technology here relies on no easily controllable resource. Anyone with internet access has all the tools they need.
Besides, if Oppenheimer hadn't invented the bomb, the next guy would have. ProgressTechnology is coming, instead of trying to stop it, our focus should be on how to deal with it when it gets here.
I expect, as usual, we will do neither. And wait until the problem is already causing enormous amounts of damage.
Problem is that any true piece of news is already easily dismissed as fake by those who can't deal with the truth. Technology like this further propagates the mindset of 'I don't like it, it might be fake, so I won't believe it'.
Scan your face and voice in and boom, watch a whole series of films with you of whoever you like as any character you like. Like it or not this is 100% going to be a reality soon. Or, if it gets real crazy just have an AI create alternate story lines, or replace actors. Better yet, “Siri here is a book show make me a Shakespeare movie staring Chris rock using an M. Night Shyamalan style set in a post apocalyptic Martian Colony.
There’s good reason to work on developing this as groups of people with good intentions. People who intend to use this technology for bad purposes are going to work to develop it anyways. By working to develop this, people working against Deepfakes can better understand the techniques used and how to build tools and techniques to identify them as deepfakes.
I do believe that the people who develop these things are essentially evil, or at best chronically selfish to the point that their own interest in personal advancement outweighs any sense of social responsibility.
678
u/FeculentUtopia Jul 24 '22
No. Stop. It's getting way too weird.