The HBO hackers strike again — this time taking control of the network’s official social media accounts.
According to tweets posted on HBO’s Twitter, a group identifying themselves as OurMine breached HBO’s main account, claiming to be “testing” the security and encouraging the network to reach out for an upgrade.
The tweets were deleted shortly after being posted, and there doesn’t seem to have been any severe damage to the company this time around.
According to The New York Times, Twitter accounts for some of the network’s most popular shows like Game of Thrones and Girls were also hacked.
OurMine, the group responsible for the breach, is infamous for targeting social media, and has successfully hacked Twitter accounts of Netflix, the WWE, and Marvel, to name a few.
However, this hack is just one of many security troubles HBO has had to face in recent history. Last month, hackers successfully breached HBO’s security to gather 1.5 terabytes of data. Also, the network has endured weeks of leaked television episodes, scripts, and other significant data — the latest including everything from West World Season 2 shooting schedules to 27 separate Game of Thrones Season 7 “shooting [diaries].”
To make matters worse, just earlier this week on Wednesday, HBO learned that Game of Thrones episode 6, which is due to air this Sunday, had leaked online, and this time that leak appears to be HBO’s own fault. (HBO Spain accidentally made the episode available to subscribers for an hour before it was removed.)
While HBO’s latest Twitter breach isn’t necessarily related to the massive data dumps that have been occurring, the timing certainly isn’t great.
Though no financial action appears to have been taken yet to prevent additional leaks, HBO should probably figure out how to get its sh*t together soon.
Content warning: This post contains racial slurs, homophobic language and very graphic depictions of racism and violence.
If you were just paying attention to press releases this week it’d be easy to believe that tech companies are winning the war on hate. Responding to the violence in Charlottesville, Mark Zuckerberg solemnly reflected that there is “no place for hate in our community.” Snapchat announced that hate speech “will never be tolerated” on its platform. YouTube reassured us that helpful tools are on the way. Tech companies fled Trump’s dual business councils to protest his claim that some white supremacists are “very fine people.”
In other headlines, a coalition of web providers made a controversial and unprecedented choice to yank their services out from under the Daily Stormer, a white supremacist news site. Days later, Cloudflare abandoned the site to the whims of whoever feels like DDoSing it. Those decisions, part of the “no platforming” philosophy which would deny hate speech purveyors a place to assemble and share their views, will likely have many reverberations in the days to come. For now, some things remain very much the same.
Unfortunately, while this week’s burst of industry energy might suggest otherwise, hate groups are alive and well, making little if any effort to conceal their presence on all of the major social networks. Whether it’s 4chan or Facebook, if you go looking for hate online, you’ll find it. Dredging up racist, anti-Semitic content often in seeming violation of a company’s stated policy takes seconds — trust me, I went looking.
On something like Facebook, hate festers just under the paper-thin layer between a user’s social sphere and the platform at large. On a network like Twitter, it’s right on the surface, bobbing unpleasantly along down the stream with dog photos and journalist chatter. For anyone surprised about the terrible events that unfolded in Charlottesville: You can find hate anywhere you look and you don’t have to look very hard.
I took a grim tour around some of the major social sites into which we sink our hours to see what just a little bit of casual searching could find — yet algorithms often can’t (or won’t). Again, this content is graphic and disturbing, but pretending it isn’t there won’t make it go away.
On Facebook, white supremacist memes thrive, even in wide-open, public communities. Though plenty of hate just sits out in the open, some users skirt detection by using a kind of unsearchable, far-right code language. Facebook might pick up on the anti-Semitic slur “kike,” but by swapping that for “kayak” the content flies under the radar. I was surprised to see that surrounding words in multiple parentheses, also called an “echo,” remains common practice to denote something or someone as Jewish. These symbols were established as part of the shallowly submerged white supremacy lexicon more than a year ago.
References to 1488 also remain common, where 14 is a nod to the “14 words” or “We must secure the existence of our people and a future for white children,” a popular mantra with white supremacists and white nationalists. The double 88 is usually a nod to the 8th letter of the alphabet, or “heil Hitler.”
Small waves of white supremacist memes crest and fall, and much like Facebook’s fake news problem, each wave has another set right behind it and there are many oceans. When I spent some time looking through these communities this week, a particularly popular meme remixed the incredible violence of a counter-protester rolling off of a now infamous ash-gray Dodge Charger with a broad array of anti-black racist memes, some of them drawing from popular mainstream memes, like “the floor is” joke. Another pictured George Washington driving the Charger through the crowd.
One public community I found easily hosted a live stream of Saturday’s white supremacist rally in Charlottesville, the full video shot from the perspective of one of the torch-bearing attendees. It felt surprising that so much of this content was just sitting right out in the open on a social network that connects faces to names.
Following Charlottesville, Facebook cracked down, removing a slew of white supremacist and white nationalist pages. Among them: Right Winged Knight, Right Wing Death Squad, Awakening Red Pill, Physical Removal, Genuine Donald Trump, Awakened Masses, White Nationalists United, Vanguard America, Radical Agenda: Common Sense Extremism and the personal page of Chris Cantwell. Many, many others remain as Facebook continues to rely on users flagging content themselves — a deeply flawed method that’s proven far more effective as a tool for harassing LGBTQ users and black activists than ridding the platform of hate.
In his statement on Wednesday, Zuckerberg did not meaningfully clarify how Facebook will determine what stays on its platform and what goes. Though he noted that “when someone tries to silence others or attacks them based on who they are or what they believe, that hurts us all and is unacceptable,” it does not appear to be unacceptable on Facebook.
Asked how its policy might be evolving, Facebook told me that it does not tolerate hate speech or posts praising acts of violence or hate groups on its platform. This policy, like all policies, is open to interpretation and it’s possible that interpretation could shift further over time.
In spite of Reddit’s mostly hands-off policy and reliance on subreddit-specific moderators, racism on Reddit often takes quirkier forms meant to avoid potential detection. In true Reddit style, overtly racist posts and comments are often played off as self-parodies, draping a thin layer of self-referential humor over what is usually just outright white supremacy. On one thread, users enthusiastically counted up from the number 1,488,000. On subreddits like /r/greentext, users post screencaps of posts from 4chan, host of some popular far right and white supremacist communities. They’re careful not to post links to 4chan itself and by screencapping they can avoid searchable text while still replicating most of the content.
In late 2015, Reddit rid itself of some popular openly white supremacist subcommunities like /r/coontown during a prominent sweep, but remarkably, pages like /r/blackpeoplehate live on. Reddit now classifies its most objectionable content as under “quarantine” and requires a verified email address to access it. Like YouTube, which took a similar approach of walling off some content, Reddit “will generate no revenue, including ads or Reddit Gold,” from these subreddits. They live on in a state of partial suspended animation.
Following the violence in Charlottesville, Reddit told me that it banned /r/physical_removal for “a violation of our content policy, specifically, the posting of content that incites violence.” The company appears responsive to user-generated campaigns when they draw sufficient attention to an issue, which appears to be the goal of /r/AgainstHateSubreddits, a compendium of Redditor-reported hate speech.
Initially, YouTube’s search made finding white supremacist stuff kind of hard. Given Google’s web search prowess it makes sense that the company would do a better job of burying objectionable content than a site like Facebook, but it wasn’t buried very deep. After a few searches didn’t turn up much, I struck Nazis on a video that prominently displayed a 1488 with a slew of links to the Daily Stormer.
Because it’s an entertainment site as much as a social network, many of my search results were home-brewed music videos depicting Nazi imagery with little or no context. A cursory glance at the user names and links was the only overt hint, with, again, many, many 1488s. Some more narrative racism came with disclaimers that the content was satire or just a joke.
Elsewhere, content drawn directly from 4chan’s infamous far right hub /pol/ (short for “politically incorrect”) was repurposed on a more mainstream platform. Because YouTube, like many of these sites, offers recommended content related to what you’re viewing, stumbling onto a little bit of white supremacy opens up a cascading slide of swastikas and racial epithets. Just a few clicks away from a music video declaring whites the master race I ran into a video created by “fashygamer1488” with the following text:
“Hey goys, its [me] here with another video, please write ur comments below, no (((jews))) or googles allowed (Google is a secret alt-right codeword that means the N word lol)…”
Again, racial slurs are traded for common, unsearchable words to keep the content just barely underground.
In June, YouTube followed Reddit’s example, creating a separate class of objectionable content that it would no longer monetize. This followed a corporate outcry from brands concerned that their ads were being served along with videos containing hate speech. In just a little bit of time spent browsing YouTube’s white supremacist content, I did not run into anything that set this content apart from the rest of its videos, though YouTube has said that feature is coming “soon” and that the “videos will remain on YouTube behind an interstitial, won’t be recommended, won’t be monetized, and won’t have key features including comments, suggested videos, and likes.”
For now, the suggestion engine hums along, pointing me toward a selection of Hitler youth haircut instructional videos.
Twitter is more responsive as a search engine than something like Facebook, but the search results are often messily curated. My first search for 1488 quickly pulls up tweets like a picture of a white, blue-eyed baby with the text “14 words” and a photo of Hitler. In other tweets, users with neo-Nazi black sun icons and hybrid Trump/Hitler background images call each other “fags” over who is and is not “boomerposting” (i.e. tweeting like a baby boomer).
Unsurprisingly, Twitter has it all. White supremacist demagogue Richard Spencer trying to remain relevant while his peers accuse him of being a Jew. Quotes hailing Trump’s off-the-rails presser that defended some white supremacists as “fine people.” Racist code words that reverse virtue-signal hate to anyone looking for a like-minded follow. Jokes about cars caked in Photoshopped blood. All of it sends the same message.
On Twitter, there is a lot, lot, lot of this content. It starts to run together.
Tech at a crossroads
These major platforms offer a taste of the toxicity flowing through mainstream social networks, but there are many others. After incubating this kind of stuff for ages, gaming chat platform Discord just finished a major purge. Tumblr, Instagram and Snapchat are fighting the same fight and it’s not clear they’re winning. Meanwhile, far right offshoots like Gab are specifically designed with sustainable white supremacy in mind. The absolute ubiquity of Nazi insignia, Stormfront links and shockingly violent memes would appear to undercut objections by the extreme right that their speech is being suppressed with much success at all.
Depending on how you use the internet, the fact that this stuff is so easy to find on major social networks could range anywhere from shocking to wholly unsurprising. But the truth is that most of us shy away from looking at it. For anyone who isn’t the target demographic, all of this hate is ugly and exhausting. We’d rather just rest easy knowing that tech companies are working on it and they’d rather we didn’t haul up more of this stuff — they’re working on it.
As we can see from tech ratcheting up its response following Charlottesville, no policy is set in stone. While companies often point users to policies around what does and doesn’t fly on their platforms, ultimately the decision to ban content is a subjective response to getting too much heat. Given that willingness to bend to public sentiment, corporate pressure and user-driven anti-hate campaigns are proving themselves to be powerful tools, even if it’s not clear where exactly to draw the line. Racial slurs? Nazi insignia? Overt threats of specific violence? For tech, the coming weeks will be a bellwether.
Anywhere you go, white supremacist content has a foothold if not an entire underground compound bedecked in red and black — one that remains even after the Charlottesville backlash. All one needs to do is look. Whether tech companies choose to see is a different matter altogether.
President Donald Trump has had an especially hard time convincing Americans he’s not racist this week.
What with blaming “many sides” and the “alt-left” after this weekend’s deadly violence in Charlottesville — despite white supremacists and neo-Nazis being at the heart of the protests against the removal of a Confederate statue — he’s making it pretty hard to believe otherwise.
Trump received a great deal of backlash for refusing to directly condemn white supremacists, the KKK, and neo-Nazis until Monday, and that statement was undermined by his off-the-rails press conference a day later that once again blamed “both sides,” emboldening the KKK and white supremacists.
Trump has yet to denounce the hate groups that brought fear into Charlottesville’s streets, beginning with a racist, torch-lit march Friday night, on Twitter, his messaging platform of choice. Holding back on Twitter insults seems pretty unlike Donny Boy.
In fact, after searching through the depths of the Trump Twitter Archive, it’s clear Trump has no problem publicly decrying things and people as racist — so long as they’re not Nazis or white supremacists.
While the archive shows many accusations from Trump, it also shows some of the many who’ve called him racist. Over the years big names like David Letterman, Al Sharpton, Donny Deutsch, and Mitt Romney have called the president racist — but he continues to assure us all he is not.
Don King, and so many other African Americans who know me well and endorsed me, would not have done so if they thought I was a racist!
The events that unfolded this weekend in Charlottesville, Virginia have sparked reactions from all corners of the internet. But one Twitter user, former President Barack Obama, had a response that resonated with millions.
Sadly, this common-sense tweet stands in direct opposition to a shocking number of people who side with the Alt-Right, including current President Donald Trump.
But what’s more interesting, perhaps, is the lens that Twitter offers us through which to view the world. In 2014, the most retweeted and liked tweets on the platform included Ellen Degeneres’ Oscars selfie.
It was followed by Justin Bieber’s… this:
YOU ARE ALL WORTHY NO MATTER WHAT ANYONE SAYS >> BE STRONG GOD IS WITH US ALL> MY BELIEBERS CHANGED MY LIFE> I WILL FOREVER BE GRATEFUL
To be fair, all hope is not lost. The #NuggsForCarter kid still holds the record for most retweets this year, at 3.42 million. But truthfully, that doesn’t offer much solace.
It’s depressing AF that the most-liked tweets of the year, thus far, are in reaction to terrorism and xenophobia, and say something that we should all inherently know: that all people are created equal.
Now, just days after the former president responded to the tragic violence with a powerful Nelson Mandela quote, his tweet has become the most-liked in history.
The record-breaking tweet in the series of three includes an image of Obama smiling through a window at four children, accompanied by Mandela’s words, “No one is born hating another person because of the color of his skin or his background or his religion…”
The tweet quickly gained over 2 million likes by Tuesday afternoon, and on Wednesday a spokesperson from Twitter confirmed to Mashable it had become the second most liked tweet of all time.
Then on Tuesday night at approximately 10:07 p.m. ET, Obama’s tweet catapulted to the top of the Twitter totem, setting a new all-time record. The tweet netted an impressive 2,705,809 likes (and rising) at the time of writing.
The second most-liked tweet now belongs to Ariana Grande, whose 2017 after the deadly Manchester bombing attack at her concert received 2,703,448 likes. And of course, the third most-liked tweet in history goes to Ellen DeGeneres’ from 2014, with 2,418,912 likes.