Social media companies in Germany face fines of up to 50m euros if they fail to remove obviously illegal content in time. From October, Facebook, YouTube, and other sites with more that two million users in Germany must take down posts containing hate
speech or other criminal material within 24 hours. Content that is not obviously unlawful must be assessed within seven days.
Failure to comply will result in a 5m euro penalty, which could rise to 50m euros depending on the severity of the offence.
Facebook responded in a statement:
We believe the best solutions will be found when government, civil society and industry work together and that this law as it stands now will not improve efforts to tackle this
important societal problem.
German MPs voted in favour of the Netzwerkdurchsetzungsgesetz (NetzDG) law after months of deliberation, on the last legislative day before the Bundestag's summer break.
Opponents responded the
tight time limits are unrealistic, and will lead to accidental censorship as technology companies err on the side of caution and delete ambiguous posts to avoid paying penalties.
The bill has faced criticism from human right's campaigners. Many of
the violations covered by the bill are highly dependent on context, context which platforms are in no position to assess, wrote the UN Special Rapporteur to the High Commissioner for Human Rights, David Kaye. He added that the obligations placed upon
private companies to regulate and take down content raises concern with respect to freedom of expression.
The law may still be chllenged in Brussels, where campaigners have claimed it breaches EU laws.
Facebook is launching a UK initiative to train and fund local organisations it hopes will combat extremism and hate speech. The UK Online Civil Courage Initiative's initial partners include Imams Online and the Jo Cox Foundation.
The recent terror attacks in London and Manchester - like violence anywhere - are absolutely heartbreaking. No-one should have to live in fear of terrorism - and we all have a
part to play in stopping violent extremism from spreading. We know we have more to do - but through our platform, our partners and our community we will continue to learn to keep violence and extremism off Facebook.
Last week Facebook
outlined its technical measures to remove terrorist-related content from its site. The company told the BBC it was using artificial intelligence to spot images, videos and text related to terrorism as well as clusters of fake accounts.
explained that it was aiming to detect terrorist content immediately as it is posted and before other Facebook users see it. If someone tries to upload a terrorist photo or video, the systems look to see if this matches previous known extremist content
to stop it going up in the first place.
A second area is experimenting with AI to understand text that might be advocating terrorism. This is analysing text previously removed for praising or supporting a group such as IS and trying to work out
text-based signals that such content may be terrorist propaganda.
The company says it is also using algorithms to detect clusters of accounts or images relating to support for terrorism. This will involve looking for signals such as whether an
account is friends with a high number of accounts that have been disabled for supporting terrorism. The company also says it is working on ways to keep pace with repeat offenders who create accounts just to post terrorist material and look for ways of
circumventing existing systems and controls.
Facebook has previously announced it is adding 3,000 employees to review content flagged by users. But it also says that already more than half of the accounts that it removes for supporting terrorism
are ones that it finds itself. Facebook says it has also grown its team of specialists so that it now has 150 people working on counter-terrorism specifically, including academic experts on counterterrorism, former prosecutors, former law
enforcement agents and analysts, and engineers.
One of the major challenges in automating the process is the risk of taking down material relating to terrorism but not actually supporting it - such as news articles referring to an IS propaganda
video that might feature its text or images. An image relating to terrorism - such as an IS member waving a flag - can be used to glorify an act in one context or be used as part of a counter-extremism campaign in another.
German authorities want the right to look at private messages on services such as WhatsApp to try and prevent terrorism. Ministers have also agreed to lower the age limit for fingerprinting minors to six from 14 for asylum seekers.
central government and federal states said encrypted messaging services, such as WhatsApp and Signal, allow militants and criminals to evade traditional surveillance. We can't allow there to be areas that are practically outside the law, interior
minister Thomas de Maiziere told reporters.
Among the options Germany is considering is source telecom surveillance, where authorities install software on phones to relay messages before they are encrypted. That is now illegal.
also planning laws to make it easier to monitor encrypted messages as well as building out a linked network of cameras and other equipment to read vehicle licence plates.
In response to recent boycotts by high profile advertisers, YouTube has clarified its censorship rules to enable video-makers to know which content it considers to be advertiser-friendly.
In a blog post, the video-sharing website said it would not
allow adverts to appear alongside hateful or discriminatory content. It will also refuse to place ads next to videos using gratuitously disrespectful language that shames or insults an individual or group. The guidelines also discourage film-makers from
making inappropriate parody videos using popular family entertainment characters.
YouTube has detailed new censorship rules in a blog post:
Hateful content: Content that promotes discrimination or disparages or humiliates an individual or group of people on the basis of the individual's or group's race, ethnicity, or ethnic origin, nationality, religion,
disability, age, veteran status, sexual orientation, gender identity, or other characteristic associated with systematic discrimination or marginalization.
Inappropriate use of family entertainment characters: Content
that depicts family entertainment characters engaged in violent, sexual, vile, or otherwise inappropriate behavior, even if done for comedic or satirical purposes.
Incendiary and demeaning content: Content that is
gratuitously incendiary, inflammatory, or demeaning. For example, video content that uses gratuitously disrespectful language that shames or insults an individual or group.
However, the announcement has met with some criticism from video makers. Captain Sauce, pointed out that the algorithm used to detect whether a video may contain inappropriate content was not perfect.
Whilst Eugenia Loli pointed out that
mainstream news networks often post inflammatory studio debates that could be judged incendiary and demeaning, while music videos often pushed the boundaries of sexually-explicit content, but these still carried advertisements. He wrote:
Why punish the little guy, but not the big networks? This is a double standard.