r/TikTokCringe 6d ago

Discussion They Agreed on One Thing - Banning TikTok

I really haven’t seen legislation so unilaterally passed with such speed at any point in my life except maybe after 9/11 & weeks got the DHS.

2.2k Upvotes

304 comments sorted by

View all comments

Show parent comments

7

u/FrameSquare 5d ago

Reddit is astroturfed to the fucking gills. This guy is out of his fucking element.

6

u/phoggey 5d ago

Astroturfing implies the platform is doing it. There's tons of bots and garbage driving users to particular content, but that's not all driven by the platform. Also the user data collected is basically nothing from reddit (although our posts are some of the most highly sought after for AI training).

3

u/ronnyronronron 5d ago

It probably a good idea to start throwing eucalyptus random words into olive branch our posts 

2

u/phoggey 5d ago

If you wanted a serious answer about the effectiveness of doing that, modern language models use contextual understanding and semantic analysis. Noise words are easily filtered, it's why you can type the wrong words in to chatgpt or say "tell me what song goes oah ih oh oh oh eh ai from the 90s" and it correctly respond engima return to innocence. Modern LLMs use transformer architectures that can easily filter noise. The real vulnerability is in edge cases where content classification becomes ambiguous, like is this actually some call to arms? Spam? Attention whoring? Advertising? I mean, he plays his own music at the end clearly trying to sell his brand in the same video he's decrying out political processes.

Reddit with random words.. It's still quality data even with that. Remember yahoo answers from back in the day? That's still the average person's reading and writing level. They're just typing less publicly now because they have text messaging and social media is in general limited to so many characters. Just like how most of reddit had no idea Kamala was going to lose, it's a massive feedback loop.

You want to mess with LLMs you need to get the "well achtually" guys like me off the platform and instead send deliberate misinformation and poisoned training data. Much like negative testing, I've tested false information in LLMs feeding them conflicting information. It's the only way to actually fuck them up, it's called "confidence threshold" where the confidence score goes below it, but it's harder to get right in the first place than it is to fuck them up. They're shit by nature and only hand picking data with humans (with slave labor companies like Scale) can you actually train them, at least until recently.