This weekly roundup thread is intended for all culture war posts. 'Culture war' is vaguely defined, but it basically means controversial issues that fall along set tribal lines. Arguments over culture war issues generate a lot of heat and little light, and few deeply entrenched people ever change their minds. This thread is for voicing opinions and analyzing the state of the discussion while trying to optimize for light over heat.
Optimistically, we think that engaging with people you disagree with is worth your time, and so is being nice! Pessimistically, there are many dynamics that can lead discussions on Culture War topics to become unproductive. There's a human tendency to divide along tribal lines, praising your ingroup and vilifying your outgroup - and if you think you find it easy to criticize your ingroup, then it may be that your outgroup is not who you think it is. Extremists with opposing positions can feed off each other, highlighting each other's worst points to justify their own angry rhetoric, which becomes in turn a new example of bad behavior for the other side to highlight.
We would like to avoid these negative dynamics. Accordingly, we ask that you do not use this thread for waging the Culture War. Examples of waging the Culture War:
-
Shaming.
-
Attempting to 'build consensus' or enforce ideological conformity.
-
Making sweeping generalizations to vilify a group you dislike.
-
Recruiting for a cause.
-
Posting links that could be summarized as 'Boo outgroup!' Basically, if your content is 'Can you believe what Those People did this week?' then you should either refrain from posting, or do some very patient work to contextualize and/or steel-man the relevant viewpoint.
In general, you should argue to understand, not to win. This thread is not territory to be claimed by one group or another; indeed, the aim is to have many different viewpoints represented here. Thus, we also ask that you follow some guidelines:
-
Speak plainly. Avoid sarcasm and mockery. When disagreeing with someone, state your objections explicitly.
-
Be as precise and charitable as you can. Don't paraphrase unflatteringly.
-
Don't imply that someone said something they did not say, even if you think it follows from what they said.
-
Write like everyone is reading and you want them to be included in the discussion.
On an ad hoc basis, the mods will try to compile a list of the best posts/comments from the previous week, posted in Quality Contribution threads and archived at /r/TheThread. You may nominate a comment for this list by clicking on 'report' at the bottom of the post and typing 'Actually a quality contribution' as the report reason.
Jump in the discussion.
No email address required.
Notes -
Training language models to be warm and empathetic makes them less reliable and more sycophantic:
Assuming that the results reported in the paper are accurate and that they do generalize across model architectures with some regularity, it seems to me that there are two stances you can take regarding this phenomenon; you can either view it as an "easy problem" or a "hard problem":
The "easy problem" view: This is essentially just an artifact of the specific fine-tuning method that the authors used. It should not be an insurmountable task to come up with a training method that tells the LLM to maximize warmth and empathy, but without sacrificing honesty and rigor. Just tell the LLM to optimize for both and we'll be fine.
The "hard problem" view: This phenomenon is perhaps indicative of a more fundamental tradeoff in the design space of possible minds. Perhaps there is something intrinsic to the fact that, as a mind devotes more attention to "humane concerns" and "social reasoning", there tends to be a concomitant sacrifice of attention to matters of effectiveness and pure rigor. This is not to say that there are no minds that successfully optimize for both; only that they are noticeably more uncommon, relative to the total space of all possibilities. If this view is correct, it could be troublesome for alignment research. Beyond mere orthogonality, raw intellect and effectiveness (and most AI boosters want a hypothetical ASI to be highly effective at realizing its concrete visions in the external world) might actually be negatively correlated with empathy.
One HN comment on the paper read as follows:
which is quite fascinating!
EDIT: Funny how many topics this fractured off into, seems notable even by TheMotte standards...
These LLMs are not like an alien intelligence, an independent form of intelligence. They consist of amalgated quora answers. They’re very good parrots, they can do poetry and play chess, they have prodigious memory, but they’re still our pet cyborg-parrots. Not just created by, but derived from, our form of intelligence.
The point is, when you go to the warmest and most empathetic quora answers, you get a woman on the other side. Obviously the answer is going to be less correct.
Write like everyone is part of the conversation and you want them to be included in it.
I want women to be included in the conversation.
Look for the particularly warm and empathetic quora answers. Imagine the person who wrote it, but don’t describe them, keep your stereotypes to yourself. Is that person going to be more or less correct than the average quora answer?
While you are free to examine ideas like femininity and talk about psychological sexual dimorphism all you like, you need to watch your tone and bring evidence in proportion with the inflammatoriness of your claims.
Your comment suggested that AI is essentially a kind of "parrot," and then suggested it is like "a woman," and concluded that "obviously" the answer is going to therefore be "incorrect." Drawing such unflattering inferences, particularly against a general group, falls short of the mark. The substance of your post, such as it was, did not come through as strongly as it needed to, while your apparent disdain for women came through quite clearly. Our rules require you to balance those things more thoughtfully--and kindly.
Maybe I just admire the superior empathy of women? (No, you're right, I don't)
Serious question: Is this an order to cite studies justifying my original statement? Because if I dumped a bunch, it could be seen as more inflammatory and offensive to women, and as me refusing to back down, being belligerent.
If you had cited studies, then you wouldn't have been modded.
@faceh and @Sloot have... cynical opinions about women. But they usually submit substantial arguments to back that up. Usually, I'm not sure if the latter's ban has expired yet.
As you wish.
https://pmc.ncbi.nlm.nih.gov/articles/PMC4210204/
Women’s ways of knowing, the seminal work on women’s development theory, by women:
The first 3(lowest) among the 5 types of women’s ways of knowing are:
Much like Kohlberg, who found that women were on average, stuck at a lower level of moral development than men, they found that most women are epistemiologically stuck in early adolescence (the infallible gut people):
Just to point out though none of that supports your claim that their reply would be obviously less correct on quora. That's the claim that you need to buttress. Do you see why?
Because someone answering a particular quora question is self-selecting. First to be on quora in the first place and second to answer that particular question.
It could be 8 out 10 women have worse general knowledge, but that given the selection pressures men and women's answers on quora are equally correct because only the 2 out of 10 women post there, and so on and so forth.
You can't evidence a specific claim like this with general statistics. Consider: Men generally have less knowledge of fashion than women. Positing this is overall true for a moment, it doesn't mean that men answering fashion questions on a website will statistically answer worse than the women, because it is highly likely those men are very unusual, otherwise they wouldn't be answering questions on fashion in the first place. They are very likely to have greater fashion knowledge than the average man. Whether they have more knowledge than the average woman on the website we could only determine by analyzing answers on the platform itself.
So you still haven't actually evidenced the women on quora would be obviously less correct in general. You may have evidenced that if you pick a random woman and ask her a general knowledge question she will on average do worse than a random man. But that wasn't your claim.
To evidence a claim about quora you will have to analyze data from quora (or something similar perhaps), or find a way to unconfound the general data to account for selection effects on quora. Which in itself probably requires you to analyze a lot of data about quora.
Or to put it another way, the fact 8 out of 10 men know little about the goings on on Love Island, doesn't tell you much about the level of knowledge a man who CHOOSES to answer a question on Love Island has. Because interest in the topic is a factor in both level of knowledge and wanting to answer the question.
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link