I don’t have a quip, just a sorrowful head shake that I somehow got in the shitty timeline.
As someone who’s autistic, fuck this with a fucking chainsaw lmfao.
> Researchers ultimately concluded that “a constructive strategy for identifying the violation of social norms is to focus on a limited set of social emotions signaling the violation,” namely guilt and shame. In other words, the scientists wanted to use AI to understand when a mobile user might be feeling bad about something they’ve done. To do this, they generated their own “synthetic data” via GPT-3, then leveraged zero-shot text classification to train predictive models that could “automatically identify social emotions” in that data. The hope, they say, is that this model of analysis can be pivoted to automatically scan text histories for signs of misbehavior.
Lemme get this straight: DARPA researches fabricated a series of words that signaled emotional states. And then, they, the DARPA researchers classified the series of words with the emotional states for the AI to train on (zero-shot classification). And then they hope to leverage the trained AI to identify “social emotions”?
Everything about this is fucking stupid.
The GPT-3 prompt could’ve been: “What are some sentences a shameful socialist/conservative/anarchist/terrorist/etc protestor/litterer/murderer/liar/etc might use?”, implicitly connecting shame a particular ideology. As such, social emotions signals more emotions by their method of generation and classification.
Suddenly, some random person is being targeted for having fucked up and they’re like, “Wtf did I do? Yes, I did shoplift from Target, but it was like a $20 shirt because my job at Wal-Mart makes me use food stamps to make ends meet. Fuck off!”
The AI automatically detects another violation of social norms.
And you’re like, “That’s an edge case…”. Yeah, sure, but it’s DARPA, we’re talking about here. That should be enough said.
It sounds like they’re focusing on the shame associated, which leads to the irony of they’ll find the awkward and uncomfortable ones but not the ardent. That sounds unwise
The worst part is, ChatGPT cannot generate anything new. It’s pre-trained, which is the P in the name.
It can only recombine the training data into forms that sort of match the training data. So, if the training data is garbage, the output will be more garbage.
And this garbage in garbage out is going to be used to harm real people.
In addition, ChatGPT lies. It hallucinates shit that is provably false, because that’s what it’s generated text needs to look like to match the training data.
So it will likely lead to a bunch of false positives, because the positive response better matches the training data.
Coming next, is a social credit score.
That already exists, and did for a long time. It just comes in multiple, disguised forms.
Because Project Insight/Pre-Crime weren’t cautionary tales, they were defense proposals.
Evergreen:
So like 100 percent of the internet? Mostly bots anyway. Everything is so dumb.
Loving this dystopian hellhole we’ve created!
Great another tool that can be used against women and minorities. With all that far right authoritarianism on the rise I can’t wait for the ai to flag me as a raging homosexual so I can end up in a labor camp or dead
Lemmings will support this as long as it is against:
- Republicans
- pedophiles
- Tr*mp supporters
- anti-vax
- climate change deniers
- racists
- N***s
Can we not say Nazi?
You’re literally the only commenter here that implies any support for this.
u/nxfsi seems like as good as any to block.
good idea. thanks.