Are tech companies really biased against conservatives?

Cato Institute has performed a review of the research that exists and came out in favor of the tech companies:

https://www.cato.org/publications/policy-analysis/why-government-should-not-regulate-content-moderation-social-media#full

Professionally, I haven’t worked on this issue, but I have worked as a developer for big tech, know who all would be working on this stuff, and researched all of the technologies in use and have thought a fair amount about this issue. I believe that I can give a fairly factual answer to the question.

One thing to note is that a majority of the people who will have developed the tools for automated moderation are Chinese and Indian. They’re just trying to achieve whatever policy aims have been handed to them by Legal.

Legal will tell them to do something very specific like, “Remove all content that incites violence.”

The way that this will be done is that they’ll create two sets of data, 1) a large body of general content from millions of users and 2) moderated content by human moderators that was flagged as having content that incites violence.

From there, they’ll use various techniques to try and make a programmatic determination of what words, phrases, etc. can be detected that differentiate those posts from the general set and make them be a part of the moderated content.

In essence, the AI (done well) just replicates the thinking of the original human moderators. Neither the Chinese nor Indian developers are going to fiddle with that any more than that. They have no motive to try and target any particular political entity. They’re just doing what Legal asked and trying to flag content that’s probably an incitement to violence. And while the large body of American developers might have a political motive, the way that you create these sorts of systems don’t really allow for personal politics to intervene, nor would anyone. They’re professional coders, not political spies sent in by the DNC. Thinking that they would is silly. As said, all anyone will do is take historical moderation decisions, and use that as a data set for detecting illicit posts.

So the question is: Who are the human moderators and what is their political leaning?

I believe that that answer varies, but I strongly suspect that it the general answer would be that it’s a large body of Indians, Philippinos, and stay-at-home moms in the Midwest.

These are not bastions of American Democratic Party support. And these people are vastly short of the number of people needed, so when they were generating the data that was fed into the AIs to train them, they had to moderate the entire Internet by hand and that’s completely infeasible for the number of people who were hired to perform that moderation.

It is very likely that these moderators, at that point in time, saw purely the content that was patently criminal and moderated only content that was patently criminal. They would have been receiving content that was flagged by other users or by very simple moderation algorithms that looked for very basic things like the word “bomb”. They would not have wanted to have dived into any moderating choices that took mental effort, time, or would be controversial. They have enough on their plate, and there were enough postings by people threatening to murder people that they don’t need to get involved with other content that’s more complex.

We can safely expect that the data which was used to train the bots was generated by non-partisan people and that they really only looked for patently illegal content.

The AI, even though it is doing little more than repeating what the human moderators do, can actually do more than they did because it uses math or outputs a measurable signal. The moderators gave a binary answer, allowed or barred. The AI can and will give a confidence rating, 0.1%, 20%, 76.43%, 99.94%, etc.

The (likely) Indian or Chinese developer will run the AI over a set of billions of posts and have it rate every single post on how illicit the AI believes that post to be. Then that developer will look at some posts that are at 50% confidence of containing nefarious content and decide from the sample, “Does that seem right?” If not, he’ll try 60%, 70%, 80%, and on until he finds some value that produces output that looks sufficiently questionable on average that it’s worth flagging to a moderator, while not producing so much output that the moderators would never be able to double-check it.

That developer will then continue raising the confidence rating until he finds a value (e.g. 95%) where nearly everything that the AI has output contains enough obviously illicit materials that you’re better off to just auto-ban rather than flag it to the human moderators. The poster can always file a review to get their content restored.

That would be stage one of AI moderation. Facebook and other platforms have almost certainly arrived at this point and are or were doing it in this way. There is no room for nor reason to expect bias to be involved.

Stage two of trying to moderate content, which I don’t know if anyone has done, would be to extend this process and look at all of the posts by users who later posted criminal content and analyze that for markers that might indicate a person of that mentality.

Having done this, you can start to rate users on an index of “likelihood to be a person who goes off the deep end”.

This index is a useful one for moderation because the AI is, fundamentally, stupid. The AI is really just looking for keywords and phrases that seem to be related to a desire to murder people. It’s wildly more advanced than simply looking for the literal word “bomb” but it is still, in essence, just doing that. There are lots of ways to create content that will ping the AIs radar while being completely benign.

Benign posts waste the time of the content moderators. You need to hire more of them to process the wheat from the chaff, which costs money, or you need to improve the AI to do a better job at detecting nefarious intent.

But AIs - even ones that use neural nets and stuff that should allow them to approximate human thought processes - can’t replicate what a human content moderator can do because they have no context.

If you raise a human, from birth, in a gravity-free black void, pumping food directly into its stomach, that human can have random phrases flashed through his brain and be rewarded when he squeezes his right hand for some and his left hand for others, and he’ll start picking out some patterns that allow you to start feeding arbitrary phrases through and get a decent output, but that human won’t know what those words actually mean. He won’t know that they are words. He’ll have no concept of what a word is, let alone what it refers to.

We are not taught vision. Our body takes in signals like electromagnetic wave lengths and air pressure waves and turns those signals into an “experience” and builds a way of understanding that signal. It invents the color purple out of detecting both blue and red wavelengths at the same time, despite there being no actual wavelength for purple. Red goes off into infinity in one direction and blue goes off in the other. Between red and blue are orange, yellow, and green. There is no purple between them. As said, the brain invents an imaginary way of interpreting the signals that it receives in a way that allows us to visualize and interact with that information, usefully.

A human brought up in a void, having phrases flashed through his brain, would simply interpret those phrases like we do colors. They indicate something, but they’re just a signal not a meaningful concept, not a language. To turn signals into language, you need to have context. You need someone to point to a chair and say, “That’s a chair.” You need to have experienced gravity and sat in a chair, so that you could appreciate the chair’s purpose and use. This mixture of sensations and experiences allows you to relate you visual and sensory systems together with your memory and create a conceptual object known as a “chair”. Minus all of that, though, living from birth in a giant endless blackness with no gravity, all you’ll interpret “That’s a chair” as is flickering colors. Maybe you’ll recognize some of the signal patterns as ones you have seen before in combination with other signal patterns, but that’s all that it will mean to you.

An AI is at this sort of disadvantage. It has no way of gaining any genuine understanding of what it is absorbing. It’s just looking at very simple patterns and trying to guess whether the patterns match data of type A or data of type B more closely, but in a way that is completely devoid of context. It’s just seeing flashes of color and chucking stuff to the left or right, based on what it has been trained to do. It is more effective than a simple word search - looking for the literal word “bomb” - but it is still very limited in its ability to moderate content intelligently.

The important thing to take away from that, is that it can’t get better. Unless you try to invent true AI and give it the ability to experience life and gain understanding of the real world and the sorts of contexts that humanity deals with, you’re always going to have the context gap problem. True AI, if generated, would pose ethical questions of slavery and the effort to create it is far too complicated for this purpose.

So, if you want to improve the output of your AI moderator, something like a “likelihood to be a person who goes off the deep end” index can be useful.

If I have a user who has never posted anything that seemed questionable, in their long posting history, then a single post that has an 85% confidence rating of containing illicit content is probably not illicit. The user has maybe posted a comment that discusses the history of lynching in the US, making it clear that lynching is bad, and it’s not a concern.

If, on the other hand, we have a poster whose posts have struck the same sorts of patterns as other people who later went off the deep end, and their newest post has an 85% confidence rating of including illicit content, then you will probably want to flag that post for human moderation.

With stage two, you can boost and suppress the illicit content algorithm with a secondary algorithm that takes a longer view of things and is looking for a more vague “general pattern of creepiness”.

I don’t know whether Facebook or anyone has moved into stage two, or plans to, or has had that idea. But, conceptually, they might (or might be doing something else that would have a similar outcome).

With stage two, all of this is still unbiased and still fundamentally tracks its way back to the moderation choices of human moderators, mostly based in India, the Philippines, and the Midwest. But, the odds that an extremist will have posts removed that weren’t strictly illegal becomes higher. Once you have instituted stage two, you will have reduced the amount of chaff that the human moderators are dealing with because it will have removed a lot more of the benign content and it will also auto-remove a lot of content that previously was too low on the confidence rating to moderate programmatically, beyond flagging for a human, now be removed automatically. The human moderators will move from mostly dealing with patently illegal content to, now, largely dealing with content that’s a little more questionable and context-driven and now they’ll be looking more at the TOS and less at the laws, and moderating stuff on the basis of hate speech and general bullying.

With stage two, extremists will find their stuff getting removed a lot more regularly.

But, almost certainly, that content will go away because the person is an extremist and so pinged the AIs radar as fitting a similar profile as a person who would be likely to eventually go off the deep end and post something patently criminal.

That’s not partisan, but partisans may well crowd to one side of the political spectrum. In that case, the algorithm, through purely impartial actions of foreigners with no political motives, working purely off laws restricting violence, nudity, etc. could indeed target one party over the other. But it will also target members of Earth Liberation Front more frequently than it will target members of the American Hiking Society. It’s purely looking for wackos, regardless of ideology.

So, in that respect, I would not worry about the big tech companies. Again, this isn’t a thing that I have personally worked on, but I can’t envision any reasonable way for partisan politics to have interceded, given the scale of the task and the people who would actually be doing the brunt of the work.

Where I think that the tech companies are being…dishonest or misleading, is in saying that they can’t do better and that the technology is just still a work in progress.

If I recall correctly, there’s about 1 police officer per 1000 people, in the US, and we can expect that that’s what it takes to get a fairly lawful society at the level we’re all comfortable living in.

Online, the ratio of moderators to users is probably 1 to 100,000 or 1 to 1,000,000 - though, I grant that I’m just guessing. While it may be that AI helps that to ratio to be more reasonable than it would seem, if the target should be 1:1000, it probably doesn’t help that much beyond the simple system of having other users manually flagging content (which is, in essence, just the same as when a citizen calls 911 to report a crime).

Facebook could just hire more moderators. They are not poor. And while, yes, technology can and will get better and could possibly become good enough that it would require very few or even no human moderators to manage, we are not to that point yet and the job needs doing.

The problem of the social media platforms isn’t that they’re biased - on that they’re clean - it’s that they’re trying to pretend like the Internet can’t be policed because it just has “so many people”. That’s not the way it works. If 1 in 4 people is a neurosurgeon, it doesn’t matter whether you’re looking at a group of 400 people or 4 billion people, you’ll always have one neurosurgeon for every non-neurosurgeon, eager to get their trepanning on. There aren’t too many posts or people on the Internet, you’re just trying to be cheap and not hire enough moderators for them all.

(Granted, some companies, like YouTube, are operating at a loss and so it’s unreasonable to expect them to hire more people. But, contrariwise, that simply indicates that it may be time for the government to subsidize these efforts where it is needed, or create its own policing force for the Internet and create standardized protocols that allow their force to operate, independent of platform.)

Thanks for this, Sage Rat.

Indians aren’t interested in American politics, just Indian politics. (And, in general, are fairly conservative in outlook.)

Chinese are generally apolitical. I presume that it’s not wise to be political as a citizen of China, nor do they have any reason to care about American politics, and (again) in general Chinese are more conservative in outlook than your average American.

Indians and Chinese, in large cities on the West Coast, are unlikely to encounter any hate speech directed towards them, let alone encounter it so frequently that they would become politically motivated to decide that they hate American Republicans and seek to find ways to target those people through algorithmic means. Their friends on social media are other Indians and Chinese, not crazy grandmas in the Deep South.

If you don’t care about the Bharatiya Janata Party, there’s no reason to expect an Indian to care about the Republican Party.

:dubious:

Also :dubious:

Uh, actually many Hindu, and Chinese people and their descendants in America do care, not at the rate of others, *but they also know how to check social media for those abuses. *

Most of the Indians and Chinese I know are either citizens or are on the way there. My Congressman is Indian, the mayor of my town is Chinese. There is also a substantial presence on the school board. Since many are relatively recent immigrants, they are still working their way up the political ranks.
And there is already a substantial cadre of second generation immigrants, who don’t care much for the politics of the countries they’ve never seen.
Yes, there is relatively little racial hatred around here. Or homophobia for that matter. But we all see the crap coming out of Washington.
True they are not as active as Latinos who have a longer history and who feel more directly threatened. But it’s coming.
I doubt they are targeting Republican posts on social media, but they certainly aren’t resisting the algorithmic targeting of them. But they certainly are no fans of the Republican party, which continues to die in California as their numbers increase.

As for the OP, this needs to be mentioned, when the effort gears against bigoted language that bias then goes to affect conservatives more… Now, in a more just world that should cause Republicans to stop pandering or getting in bed with the racists and bigots among their supporters, but since this is not a just world it is the tech company the one that pulls back against that capability.

“Twitter reportedly won’t use an algorithm to crack down on white supremacists because some GOP politicians could end up getting barred too”

Seems strange this is even up for debate since the whole Google memo fiasco.

The grand majority do not stay in the US. They work here for a few years and then return to their country.

The ones who don’t fit back home would, presumably, be the ones to stay.

I thought that memo was more about the differences between men and women? By the way, you might want to read this article from Wired about how that memo writer got the science wrong.
edited to add: You think one memo settles the question just because you agree with the memo?

Granted, I’ve only worked a handful of tech jobs, and it was server administration and network support, not development, but I generally found that my colleagues were unabashedly conservative. And this was in places where you’d expect people to fall more toward the left wing, such as a couple of University campuses, in the New England area, which tends to be liberal overall.

Again, it’s too small a sample size to really draw much of an inference from, but if tech workers are as conservative as what I experienced, it would be silly to think that they are censoring themselves. And indeed, social media played a big part in Donald Trump’s election, thanks to his ability to use the platforms to get his base sufficiently riled up.

The algorithm removes according to the law and the TOS. It’s their job description to enforce the TOS so of course they aren’t resisting.

Any TOS that says “no hate speech, no bullying” is very liable - at this moment in time - to disproportionately target Republicans. And any change to the TOS to remove that restriction would likely cause their platform to lose customers, because most people, left or right, don’t want to hang around a place full of bullies and racists.

Do you object to the TOS standards?

:confused: You mean, the James Damore thing where a Google employee got fired for publicly criticizing his employer’s personnel policies based on shoddy cherry-picked “scientific” arguments? What’s anti-conservative about that? Conservatives are usually pretty supportive of the principle that employers don’t owe their employees the right to publicly criticize and denigrate them, AFAICT.

That’s one way of spinning it, I suppose. Another, perhaps more honest account is that Damore provided feedback on Google’s diversity policy because Google asked for it. He then posted it in an online space Google had specifically set aside to receive such feedback. It was then shared outside that space by other people without Damore’s knowledge or consent. It then went viral, caused an international firestorm of controversy, and then Google fired him.

So why did Google fire him? Was it for using shoddy arguments? Unlikely. Damore received support from numerous psychologists and biologists who say that he actually got the science largely correct. Even if, on balance, he didn’t, it’s not like he was advocating phrenology or the miasma theory. He advocated one side of an ongoing scientific debate with good arguments on both sides. Firing is, emphatically, not a fair response to that.

It’s clear to me that he was fired because of how other employees and certain elements in the media reacted to the memo.

The reaction was furious, and utterly disproportionate to the perceived offence. From the outside, it was like watching a violent auto-immune reaction. Damore was widely castigated on social media as a ‘tech bro’ (which is hilarious, really. Watch an interview with the guy. You couldn’t imagine anyone less “bro”), an alt-righter, a fascist, and even a Nazi. Some women at Google even took sick leave, citing “emotional distress”.

While Damore’s position on gender diversity in tech isn’t conservative per se, aggressive (some might say psychotic), wildly OTT adherence to “blank slateism” is most assuredly a progressive liberal position. Given the impetus for this witch-hunt originated from within Google, it’s reasonable that, as an organisation, Google is hostile, not only to conservatives, but anyone who isn’t progressive.

The fact that a company asks for feedback does not immunize respondents against making wildly inappropriate arguments.

This is beside the point, but it was a handful of degreed individuals whose opinions aren’t supported by experimental research, and whose opinions are considered controversial if not outright garbage.

Good. People aren’t entitled to employment. Employers should fire employees who offend large numbers of their other employees. You have absolutely no right to offend people in the workplace.

The OP says you’re wrong. The argument “nature says we’re right” is the primordial conservative position. And are you suggesting that someone with psychotic views on gender diversity should share those views with his co-workers and get to keep his job? Yes. Absolutely 100% conservative. Why do you think I should work with a psychotic person who thinks my race or gender makes me less of an engineer than him?

The arguments in Damore’s memo weren’t “wildly inappropriate” or “psychotic”. They were actually fairly tepid, cautiously phrased, and largely well cited. That doesn’t mean he’s right, of course, but his position is a defensible one. That observation alone is enough to negate the rest of your post.

Damore didn’t make this argument.

You introduced the word “psychotic” in your post, not me, but thanks for clarifying. As for wildly inappropriate… of course “inappropriate” is subjective, but we are speaking in the context of employment, and employers almost universally regard it as inappropriate to send communications that offend the rest of the company based on their race and gender. This is pretty much the universal letter of the law in HR.

There is no defense of “but science” when blasting out an email suggesting that many of your co-workers shouldn’t be there because of their race or gender. Especially, ESPECIALLY when that “science” is cherry-picked, widely disputed, and far from settled.

He appealed directly to biological theories of innate intellectual ability which is a specific form of the appeal to nature.

Using Damore’s case as proof that Google is biased is silly. Every corporation the size of Google will have problems with individual employees of all spectra. The woman who organized walkouts left Google claiming retaliation. I’m sure Fox (and CNN) have had similar issues. If you have some wide-ranging stats I’m all ears.

I doubt Google has any enterprise-level policies with bias. What probably is biased is the user base; internet users on the whole tend to be younger which loosely correlates to more liberal. In machine learning environments the users have a very real influence on future results. For example, if users do an image search in Google for “baboon” some will click on Trump’s picture and some will click on, say, Obama’s picture. Given the bias of the users Trump’s picture will rise to the top, which then further re-enforces the bias. Soon image searches for “baboon” feature Trump as the first result. This is not in any way a bias of Google but its users.

Any company with extensive ML (which is most of the big internet companies) could have similar type biases.

Um, Damore’s memo included many phrases like

These are not “tepid” or “cautiously phrased” remarks. They’re the self-expression of somebody with a massive grievance lashing out at a fundamental principle of his employer’s HR policy and trying to pretend it’s about basic human biology. I completely agree that he should be legally allowed to make such adversarial remarks, but I don’t see anything anti-conservative about his employer’s not wanting to keep him around after his adversarial remarks went viral.

Yes, but in relation to the genuinely unhinged reaction to the memo. Damore was literally at the centre of an international moral panic. All because he advanced a fairly tepid and heavily qualified “nature > nurture” argument.

The fact the memo so deeply offended so many people is just weird when you consider what’s actually in it. The reaction was unreasonable, and it’s equally unreasonable to presume Damore should’ve seen it coming. Given this, Google should’ve taken a more moderate approach. That they didn’t indicates that the executives either hold, or at the least are overly sympathetic to, the beliefs of their most militantly “blank-slatist” employees.

The memo wasn’t “blasted out”. It was submitted specifically to the online space Google had reserved for feedback. And Damore emphatically didn’t say that any of his co-workers shouldn’t be there. Not even once.

And if Damore should’ve refrained from giving his opinions because the science is “far from settled”, Google should’ve fired his critics, too. After all, if his side of the argument is far from settled then theirs is, too.

Yes, but he didn’t say that any of his co-workers didn’t deserve to be there.