Facebook has started nudging people to not only create live videos, but consume them too. An update automatically pushes more live video to your smartphone screen. A bunch of people have been seeing a message pop up on their mobile News Feed: “Feeling the winter vibes? Go Live and winter it up with friends,” the company suggests.
But with Facebook Live, exhibitionists can have a home. Why would a camgirl be on Facebook without a direct revenue model? The video above also included a link to a third-party website that requires a subscription. That would technically be in violation of Facebook’s Community Guidelines, but evidently no one reported the video.
And of course, there’s always PayPal.
New look, old problems
Combating porn and piracy clearly is far from a new issue, let alone one exclusive to Facebook. Twitter’s Periscope and (the now-defunct app) Meerkat both faced the same problem.
But Facebook looks like they’re going heavy on marketing and promoting live video, while remaining light on progress where technology and editorial curation are concerned.
People have been employing Facebook Live for provocative use cases as soon as they possibly could. For example, in May, three teens allegedly streamed sexual acts, publicly, to the network. Still, we’ve learned Facebook’s policies very much leave its users exposed to explicit content.
Facebook calls itself a tech company, but it doesn’t use software to moderate pornographic or violent videos playing live on the network (to a potential daily audience of more than 1 billion users). Instead, it relies on human viewers and human “editors”—though Facebook would prefer we call them “reviewers”—to monitor, flag and take action against content in violation its editorially-chosen policies.
Over email, a Facebook spokesperson explained the company’s stance on violent or pornographic live content: “We don’t use any systems to filter out certain videos. We are not proactively reviewing content and depend on our community to report violations.”
It does, however, have a system called Rights Manager, where copyright holders can see and flag a live video, and then, request it to be taken down.
Facebook’s actions, or rather inactions, aren’t illegal. It follows the Digital Millennium Copyright Act, which states that a rights holder must submit a takedown notice to a service provider for a video to be removed. It also allows explicit content. So does, at least for now, Facebook, as long as nobody reports it.
That’s an issue for Facebook as it pushes Live as its next big thing. It’s telling you to go live from its ads on television, on billboards, and on your News Feed. It’s even paying celebrities and publishers—including Mashable—to go live. And yet, it hasn’t prioritized quality.
Then again, it’s not that surprising. The issue follows in step with Facebook’s tendency to lack immediate quality assurance and transparency. It’s worth remembering that Live isn’t Facebook’s only product with a porn problem—remember the launch of Marketplace?
Streaming Live, sans solution
The technology to detect what’s in live video does exist, whether deployable from Facebook’s codebase or not. Other live video providers including Twitter’s Periscope and YouNow told us they employ both software and human reviewers to moderate video content.
“We use a combo of tools and humans who validate if content violates guidelines before anything is removed,” a Twitter spokesperson said.
“YouNow uses both automated and manual systems to detect undesirable content from our service, and to bring this content to the forefront so that our 24/7 moderation team may review it for possible removal,” wrote a YouNow spokesperson over email.
Twitter said it still requires a human reviewer to remove a live video from Periscope, but the tech does help for flagging. YouNow, however, will let its software remove content.
“We believe it is important to use human judgment to review content that is violating our policies. That being said, in certain circumstances where we are certain that content is in violation of our Terms of Use or community guidelines, this content may be removed prior to a human approving the removal,” a YouNow spokesperson wrote.
And yes, live video detection has already been built by third-party companies. For example, Dextro, a computer vision company that worked with Periscope (prior to being owned by Twitter) and YouNow, has software to detect objects such as guns.
Meanwhile, Facebook, one of the world’s most valuable technology giants, doesn’t use software to remove potentially terms-violating videos. According to what Facebook told us, the company relies on you and a rights holder to alert them, and then, a human to make the final judgment to remove it.
“Given the level of resources [Facebook] has, it’s surprising,” said David Luan, cofounder of Dextro. “I think what must be the case is there’s some sort of specific set of product requirements they have to get right.”
Perhaps Facebook’s process is fast — even faster than Twitter’s or YouNow’s — without the tech layer. Facebook declined to disclose how big its moderation team is.
Facebook also holds an advantage over other networks within its real name policy. While people can expose themselves or share pirated content somewhat anonymously on Periscope, Facebook makes it more difficult to create fake accounts by requiring authentic names.
Interestingly—despite not actively using tech to remove explicit live video content automatically—Facebook has repeatedly blamed technical glitches for several instances of live video being taken down. For example, the company said a “technical glitch” led to the temporary removal of a video that captured the last moments of Philando Castile after being shot by a police officer.
“It only takes one report for something to be reviewed,” Facebook said in the blog post it released after the incident with the video of Castile.
Prioritizing (profitable) product
Facebook’s built and deployed a system to address copyright infringements. In April (the same month Facebook Live became available to everyone) Facebook introduced Rights Manager, where publishers can upload their content to Facebook’s system. A system called “Audible Magic” detects video and audio within live and upload videos. If something matches, the live video can be reported and then removed.
YouTube, the older video giant, has had a similar system called Content ID, but not only does it help keep copyright holders at ease, it also helps them make money.
The Financial Times reported Wednesday that Facebook was “working to build out a copyright tracking technology” similar to the one used by YouTube. Part of this technology already exists through Audible Magic, but according to the report, Facebook aims to allow rights holders to not solely request for content to be removed but to earn revenue from licensing agreements.
Facebook declined to comment on the FT report or answer further questions about its work building a new system, because the company doesn’t comment on speculation. The company confirmed it’s dedicating more resources, such as making more hires, to address copyright issues.
A Facebook spokesperson said that outside of “always working to make improvements,” the company has nothing to announce right now.
At its current state, Facebook Live is a mess.
Beyond clear violations and questionable material, there’s little discoverability.
As a Facebook spokesperson told us—and as the content evidently shows—there aren’t any editors. That’s a far cry from Periscope, which hired an editor-in-chief, and YouNow, which also has a team that chooses editor’s picks. Meanwhile, Facebook’s the company that laid off the journalists that once helped trending news topics and, at times, selected live videos to represent them.
Source:https://mashable.com
Working as a cyber security solutions architect, Alisa focuses on application and network security. Before joining us she held a cyber security researcher positions within a variety of cyber security start-ups. She also experience in different industry domains like finance, healthcare and consumer products.