Facebook表示,其管理社交媒體網站所使用的人工智能系統現在已經足夠強大,可以自動識別網站上超過94%的仇恨言論,并且可以發現超過96%與有組織仇恨團體有關的內容。
這意味著Facebook管理其網站的能力大幅提升,在個別情況下,其人工智能系統發現違規內容的效率比一年前提升了五倍。
然而,只要Facebook繼續給政治強權人物和流行的極端主義媒體組織特殊待遇,技術上的進步根本不可能改變該公司糟糕的公共形象。
最近幾周,Facebook遭到抨擊,因為在美國總統唐納德·特朗普發布有關大選的言論之后,Facebook沒有采取更多措施阻止這些言論傳播,而特朗普的前顧問史蒂夫·班農在Facebook上傳播要求將兩位曾激怒總統的美國高級官員斬首的播客,Facebook也沒有將其封禁。
Facebook后來確實將特朗普的帖子標記為誤導性信息,比如他聲稱自己贏得大選的帖子,Facebook還在一些帖子上附上了“計票工作需要幾天或幾周才能完成”的備注。但批評者認為,Facebook應該直接將特朗普的帖子刪除或者屏蔽。另外一家社交媒體公司Twitter臨時屏蔽了特朗普競選團隊的官方賬號發布的新帖子,以及特朗普的顧問在大選之前發布的帖子。Facebook稱特朗普的帖子符合其標準政策中有關“新聞價值”的豁免條款。
Facebook CEO馬克·扎克伯格表示,班農的帖子已經被刪除,但這個右翼煽動者并沒有頻繁違反公司的規定,因此不足以從Facebook平臺上將其封殺。
Facebook首席技術官邁克·斯科洛普夫承認,增強公司的人工智能系統,使其能夠監測甚至在許多情況下可以自動屏蔽違反公司規定的內容,并不能完全解決公司面臨的有害內容問題。
斯科洛普夫說:“我在這方面沒有天真的想法。我不會說技術是所有問題的解決方案。”他表示,Facebook對于社交網絡的管理分為三個方面:能夠識別違反公司政策的內容的技術,能夠根據信息快速采取行動從而防止內容產生影響的能力,以及政策本身。他補充說,在前兩點上,技術可以有所幫助,但它無法決定政策。
公司越來越多地借助自動化系統,為15,000名人類內容審查員提供幫助。許多審查員屬于公司在全球聘請的承包商。今年,Facebook首次開始使用人工智能確定將內容交由人類審查員審查的順序,由審查員確定將其保留還是刪除。人工智能軟件會根據內容可能違反政策的嚴重性以及內容在Facebook的社交網絡上傳播的可能性,確定內容的優先順序。
斯科洛普夫表示,該系統的目的是盡量限制Facebook所說的“流行度”,這個指標大體上可以理解成會有多少用戶可能看到一條內容或者與內容互動。
該公司還迅速在其內容審查系統中部署了多種由其內部研究人員開發的尖端人工智能技術。其中有一款軟件可以翻譯100種語言,無需使用一種通用中間語言。這款軟件幫助公司的人工智能打擊仇恨言論和失實信息,尤其是以不常見的語言書寫的言論,因為這些語言的人類內容審查員遠遠不足。
斯科洛普夫表示,公司在“相似度匹配”方面取得了巨大進步。相似性匹配會確定一條新內容與因為違反Facebook政策已經被刪除的另一條內容是否基本相似。他列舉了與新冠疫情有關的失實信息的例子。當時有帖子謊稱外科醫用口罩中含有已知的致癌物,經過人類事實核查員評估之后這些帖子被刪除,第二條帖子使用了略有不同的語言和類似但并不完全相同的口罩圖片,也被人工智能系統識別出來并自動屏蔽。
他還表示,許多系統現在擴充到了“多模態”領域,可以結合文本與圖片或視頻進行分析,有時候也會結合語音。Facebook有專門的軟件用于監測具體類型的惡意內容,比如識別廣告垃圾郵件的軟件和識別仇恨言論的軟件等,同時它還推出了一個新系統“整個帖子完整性嵌入”(簡稱“WPie”),這款軟件可以識別各種各樣不同類型的政策違規情況,不需使用各類違規的大量實例進行培訓。
另外,Facebook還利用研究競賽,幫助其開發更強大的內容審查人工智能。去年,Facebook公布的競賽結果顯示,研究人員開發出可自動識別深度偽造視頻的軟件。深度偽造視頻本身也是使用機器學習技術開發的高度逼真的虛假視頻。Facebook目前的競賽,旨在找到檢測仇恨模因的最佳算法。這次競賽極具挑戰性,因為一個成功的系統需要理解模因中的圖片和文字對意義的影響,甚至需要理解模因本身不存在的大量語境。(財富中文網)
翻譯:劉進龍
審校:汪皓
Facebook表示,其管理社交媒體網站所使用的人工智能系統現在已經足夠強大,可以自動識別網站上超過94%的仇恨言論,并且可以發現超過96%與有組織仇恨團體有關的內容。
這意味著Facebook管理其網站的能力大幅提升,在個別情況下,其人工智能系統發現違規內容的效率比一年前提升了五倍。
然而,只要Facebook繼續給政治強權人物和流行的極端主義媒體組織特殊待遇,技術上的進步根本不可能改變該公司糟糕的公共形象。
最近幾周,Facebook遭到抨擊,因為在美國總統唐納德·特朗普發布有關大選的言論之后,Facebook沒有采取更多措施阻止這些言論傳播,而特朗普的前顧問史蒂夫·班農在Facebook上傳播要求將兩位曾激怒總統的美國高級官員斬首的播客,Facebook也沒有將其封禁。
Facebook后來確實將特朗普的帖子標記為誤導性信息,比如他聲稱自己贏得大選的帖子,Facebook還在一些帖子上附上了“計票工作需要幾天或幾周才能完成”的備注。但批評者認為,Facebook應該直接將特朗普的帖子刪除或者屏蔽。另外一家社交媒體公司Twitter臨時屏蔽了特朗普競選團隊的官方賬號發布的新帖子,以及特朗普的顧問在大選之前發布的帖子。Facebook稱特朗普的帖子符合其標準政策中有關“新聞價值”的豁免條款。
Facebook CEO馬克·扎克伯格表示,班農的帖子已經被刪除,但這個右翼煽動者并沒有頻繁違反公司的規定,因此不足以從Facebook平臺上將其封殺。
Facebook首席技術官邁克·斯科洛普夫承認,增強公司的人工智能系統,使其能夠監測甚至在許多情況下可以自動屏蔽違反公司規定的內容,并不能完全解決公司面臨的有害內容問題。
斯科洛普夫說:“我在這方面沒有天真的想法。我不會說技術是所有問題的解決方案。”他表示,Facebook對于社交網絡的管理分為三個方面:能夠識別違反公司政策的內容的技術,能夠根據信息快速采取行動從而防止內容產生影響的能力,以及政策本身。他補充說,在前兩點上,技術可以有所幫助,但它無法決定政策。
公司越來越多地借助自動化系統,為15,000名人類內容審查員提供幫助。許多審查員屬于公司在全球聘請的承包商。今年,Facebook首次開始使用人工智能確定將內容交由人類審查員審查的順序,由審查員確定將其保留還是刪除。人工智能軟件會根據內容可能違反政策的嚴重性以及內容在Facebook的社交網絡上傳播的可能性,確定內容的優先順序。
斯科洛普夫表示,該系統的目的是盡量限制Facebook所說的“流行度”,這個指標大體上可以理解成會有多少用戶可能看到一條內容或者與內容互動。
該公司還迅速在其內容審查系統中部署了多種由其內部研究人員開發的尖端人工智能技術。其中有一款軟件可以翻譯100種語言,無需使用一種通用中間語言。這款軟件幫助公司的人工智能打擊仇恨言論和失實信息,尤其是以不常見的語言書寫的言論,因為這些語言的人類內容審查員遠遠不足。
斯科洛普夫表示,公司在“相似度匹配”方面取得了巨大進步。相似性匹配會確定一條新內容與因為違反Facebook政策已經被刪除的另一條內容是否基本相似。他列舉了與新冠疫情有關的失實信息的例子。當時有帖子謊稱外科醫用口罩中含有已知的致癌物,經過人類事實核查員評估之后這些帖子被刪除,第二條帖子使用了略有不同的語言和類似但并不完全相同的口罩圖片,也被人工智能系統識別出來并自動屏蔽。
他還表示,許多系統現在擴充到了“多模態”領域,可以結合文本與圖片或視頻進行分析,有時候也會結合語音。Facebook有專門的軟件用于監測具體類型的惡意內容,比如識別廣告垃圾郵件的軟件和識別仇恨言論的軟件等,同時它還推出了一個新系統“整個帖子完整性嵌入”(簡稱“WPie”),這款軟件可以識別各種各樣不同類型的政策違規情況,不需使用各類違規的大量實例進行培訓。
另外,Facebook還利用研究競賽,幫助其開發更強大的內容審查人工智能。去年,Facebook公布的競賽結果顯示,研究人員開發出可自動識別深度偽造視頻的軟件。深度偽造視頻本身也是使用機器學習技術開發的高度逼真的虛假視頻。Facebook目前的競賽,旨在找到檢測仇恨模因的最佳算法。這次競賽極具挑戰性,因為一個成功的系統需要理解模因中的圖片和文字對意義的影響,甚至需要理解模因本身不存在的大量語境。(財富中文網)
翻譯:劉進龍
審校:汪皓
Facebook has revealed that the artificial intelligence systems it uses to police its social media sites are now good enough to automatically flag more than 94% of hate speech on its social media sites, as well as catching more than 96% of content linked to organized hate groups.
This represents a rapid leap in Facebook's capabilities—in some cases, these A.I. systems are five times better at catching content that violates the company's policies than they were just one year ago.
And yet this technological progress isn't likely to do much to improve Facebook's embattled public image as long as the company continues to make exceptions to its rules for powerful politicians and popular, but extremist, media organizations.
In recent weeks, Facebook has been under fire for not doing more to slow the false claims about the election made by U.S. President Donald Trump and not banning former Trump advisor Steve Bannon after he used Facebook to distribute a podcast in which he called for the beheading of two U.S. officials whose positions have sometimes angered the president.
Facebook did belatedly label some of Trump's posts, such as ones in which he said he had won the election, as misleading and appended a note saying that "ballot counting will continue for days or weeks" to some of them. But critics said it should have removed or blocked these posts completely. Rival social media company Twitter did temporarily block new posts from the official Trump campaign account as well as those from some Trump advisors during the run-up to the election. Facebook said Trump's posts fell within a "newsworthiness" exemption to its normal policies.
As for Bannon's posts, Facebook CEO Mark Zuckerberg said they had been taken down but that the rightwing firebrand had not violated the company's rules frequently enough to warrant banning him from the platform.
Mike Schroepfer, Facebook's chief technology officer, acknowledged that efforts to strengthen the company's A.I. systems so they could detect—and in many cases automatically block—content that violates the company's rules were not a complete solution to the company's problems with harmful content.
"I'm not naive about this," Schroepfer said. "I'm not saying technology is the solution to all these problems." Schroepfer said the company's efforts to police its social network rested on three legs: technology capable of identifying content that violated the company's policies, the capability to quickly act on that information to prevent that content from having an impact and the policies themselves. Technology could help with the first two of those, but could not determine the policies, he added.
The company has increasingly turned to automated systems to help augment the 15,000 human content moderators, many of them contractors, that it employs across the globe. This year for the first time, Facebook began using using A.I. to determine the order in which content is brought before these human moderators for a decision on whether it should remain up or be taken down. The software prioritizes content based on how severe the likely policy violation may be and how likely the piece of content is to spread across Facebook's social networks.
Schroepfer said that the aim of the system is to try to limit what Facebook calls "prevalence"—a metric which translates roughly into how many users might be able to see or interact with a given piece of content.
The company has moved rapidly to put several cutting-edge A.I. technologies pioneered by its own researchers into its content moderation systems. These include software that can translate between 100 languages without using a common intermediary. This has helped the company's A.I. to combat hate speech and disinformation, especially in less common languages for which it has far fewer human content moderators.
Schroepfer said the company had made big strides in "similarity matching"—which tries to determine if a new piece of content is broadly similar to another one that has already been removed for violating Facebook's policies. He gave an example of a COVID-19-related disinformation campaign—posts falsely claiming that surgical face masks contained known carcinogens—which was taken down after review by human fact-checkers and a second post that used slightly differently language and a similar, but not identical face mask image, which an A.I. system identified and was able to automatically block.
He also said that many of these systems were now "multi-modal"—able to analyze text in conjunction with images or video and sometimes also audio. And while Facebook has individual software designed to catch each specific type of malicious content—one for advertising spam and one for hate speech, for example—it also has a new system it calls Whole Post Integrity Embedding (WPie for short) that is a single piece of software that can identify a whole range of different types of policy violations, without having to be trained on a large number of examples of each violation type.
The company has also used research competitions to try to help it build better content moderation A.I. Last year, it announced the results of a contest it ran that saw researchers build software to automatically identify deepfake videos, highly-realistic looking fake videos that are themselves created with a machine learning technique. It is currently running a competition to find the best algorithms for detecting hateful memes—a difficult challenge because a successful system will need to understand how the image and text in a meme affect meaning as well as possibly understand a lot of context not found within the meme itself.