Join me @ IBOtoolbox for free.
Vlad Tverdohleb
Member Since: 7/26/2015
performance / stats
Country: Canada
Likes Received: 151
Featured Member: 1 times
Associates: 155
Wall Posts: 382
Comments Made: 62
Press Releases: 370
Videos: 0
Phone: 015144814545
Skype:     theprservices
profile visitor stats
TOTAL: 72661
are we ibo associates?
recent videos
member advertising
active associates
Whitney Jacqueline    
Last logged on: 7/21/2019

Last logged on: 7/21/2019

Velma Joseph     
Last logged on: 7/21/2019

Mark Turnbull    
Last logged on: 7/21/2019

Mike Farris    
Last logged on: 7/21/2019

Pavel Mrlian    
Last logged on: 7/21/2019

Bruno Duarte    
Last logged on: 7/21/2019

Csaba Juhasz    
Last logged on: 7/21/2019

Kenneth Duremdes    
Last logged on: 7/21/2019

Elena Garas    
Last logged on: 7/21/2019

IBOtoolbox Admin     
Last logged on: 7/21/2019

Sule Yesufu    
Last logged on: 7/21/2019

Eugenijus Sakalauskas    
Last logged on: 7/21/2019

Rix Robinson    
Last logged on: 7/21/2019

Cosmos Parris    
Last logged on: 7/21/2019

other ibo platforms
Vlad Tverdohleb   My Press Releases

Facebook Will Start Making Its Own AI Chips To Stop Violent Livestreams

Published on 6/5/2018
For additional information  Click Here

When you give people a channel to send out live videos, some will use it to broadcast the most horrendous, disgusting things imaginable. Facebook has learned this first hand; the site has become infamous for livestreams of violent acts, murder, and even suicide.

To keep these kinds of violent videos off the platform, Facebook introduced systems, both AI systems and human moderators, to scan for violent broadcasts and take them down. Gradually, the company has relied more heavily on its AI moderators.

Facebook, along with other online platforms, has been developing these violence-flagging algorithms for a while. But now Facebook has plans to develop its own AI hardware — a chip that hosts the AI. The biggest advantage of a chip? They require much less computing power, which would lead to much faster algorithms, according to Bloomberg. As it is now, the AI filters can catch violent videos in about ten minutes on average, though sometimes they stay on the site for hours. Ideally, though, Facebook would like to take those livestreams down as they’re happening. There’s no telling whether a chip like this would get them to that goal, but it would almost definitely get them closer to it.

This seems like a good investment for a company that projects a more-or-less wholesome, family-friendly social network. And if their video-catching tools do drastically improve how quickly videos of suicides and murders are flagged, some of those violent acts may not even happen since the perpetrators wouldn’t have an audience. That’s the same logic that stops the media from glorifying serial killers.

But what remains unclear is how Facebook will train its algorithm to either flag or permit violent videos posted by activists of bystanders to raise awareness of violent acts perpetrated by others. Like, say, to monitor police violence.

When Jeronimo Yanez, a then-police officer shot and killed Philando Castile in 2017, Castile’s girlfriend filmed the event on Facebook Live. Facebook couldn’t decide what to do with the footage, first deleting, then re-uploading it with graphic content warnings before ultimately removing it again. Facebook’s logic was that videos raising awareness of violence would be allowed, while those celebrating it would be deleted.

Read more here.

Member Note: To comment on this PR, simply click reply on the owners main post below.
-  Copyright 2016 IBOsocial  -            Part of the IBOtoolbox family of sites.