維諾德·科斯拉與馬克·安德森都從公司創始人轉型為投資者。上周末,兩人就通用人工智能開發是否應該開源展開辯論。通用人工智能將使機器的智能水平與人類相當。
兩人爭論的導火索是科斯拉發布了一條帖子,對OpenAI及其首席執行官薩姆·奧爾特曼大加贊揚。
科斯拉寫道:“我們從@OpenAI創立之初就認識@sama,我們完全支持他和他的公司。這些訴訟是開發通用人工智能和實現其效益的巨大干擾。”
對于科斯拉的信息,安德森指責他“游說禁止開源”人工智能研究。
安德森似乎對科斯拉支持OpenAI的立場有不同意見,因為OpenAI已經背離了最初的開源原則。自人工智能誕生以來,安德森就堅定支持開源人工智能,他認為開源人工智能是一種保障措施,可以避免少數幾家大型科技公司和政府部門掌控最前沿人工智能研究的準入。
在當前的這場論戰和以前的言論中,安德森對于一些最大的人工智能批評者提出的擔憂嗤之以鼻。安德森曾經把這些擔憂歸咎于對顛覆性和不確定性的恐懼,而不是技術本身具有惡意,他在X上重申了這種觀點。
安德森在X上發帖稱:“每一種能夠提升人類福祉的重大新技術,都會引發虛假的道德恐慌。人工智能則是最新的例子。”
另一方面,科斯拉卻透過地緣政治和國家安全的視角看待人工智能,而不是從嚴格的創業角度來看待它。
對于安德森所說的科斯拉不支持開源的言論,科斯拉回應稱開源風險過高。
他回復安德森稱:“你會將曼哈頓計劃(Manhattan Project)開源嗎?人工智能對國家安全更加重要。我們必須在人工智能領域取得勝利。這是愛國主義,而不是口號。”
科斯拉和安德森之間的辯論,涉及到薩姆·奧爾特曼、OpenAI面臨的訴訟和埃隆·馬斯克。馬斯克后來也加入了論戰。兩人在辯論中還談到了是否應該允許任何人進行任何形式的人工智能研究,或者是否應該將最先進的人工智能版本交給政府。雖然這場辯論看起來只是一群非常成功的硅谷企業家在網絡上展開的一些討論,但我們可以管中窺豹,從中看到目前圍繞開源人工智能展開的關鍵辯論。
最終,沒有任何一方希望徹底禁止開源或閉源研究。在這場關于限制開源研究的論戰中,有部分主張源于擔心限制開源研究被作為一種惡意的理由,目的是確保對在人工智能領域已經取得進展的大公司實施監管。傳奇人工智能研究人員、Meta前首席人工智能科學家楊立昆(Yann LeCun)在X上加入論戰時表達了這種觀點。
他寫道:“沒有人要求禁止閉源人工智能。但有些人正在積極游說各國政府禁止(或限制)開源人工智能。有些人以軍事和經濟安全作為理由。還有人則提到了威脅人類生存的幻想。”
硅谷知名天使投資人羅恩·康威要求領先的人工智能公司承諾“開發能夠改善生活,為人類開啟更美好未來的人工智能”。迄今為止,Meta、谷歌(Google)、微軟(Microsoft)和OpenAI都已經在他的信中簽字。
安德森引用科斯拉使用的曼哈頓項目的比喻,對OpenAI的安全保障提出了擔憂。他認為,如果沒有與曼哈頓項目同等程度的安全保障,例如“嚴格的安全審查和許可流程”、“持續的內部監控”和有“全天候武裝警衛”守護的“牢不可破的物理設施”,那么OpenAI最先進的研究就將被美國的地緣政治對手們竊取。
OpenAI并未立即回復置評請求。
但安德森似乎不只是在爭論一種觀點,而是在進行思考練習。他在回應自己的帖子時寫道:“當然,每一個假設都很荒謬。”
埃隆·馬斯克加入論戰批評OpenAI的安全措施
此時,OpenAI的聯合創始人埃隆·馬斯克加入了論戰。
馬斯克回復安德森討論OpenAI安全措施的帖子稱:“國家行為者當然很容易盜取他們的知識產權。”
科斯拉也提到了馬斯克,稱馬斯克起訴OpenAI的決定是出于“酸葡萄心理”。上周,馬斯克起訴OpenAI,指控該初創公司違反了創業協議。馬斯克認為,OpenAI與微軟的密切關系和停止將其產品開源的決定,違背了公司創立的使命。據彭博社(Bloomberg)掌握的一份備忘錄顯示,OpenAI與科斯拉的觀點類似,指責馬斯克“后悔沒有參與公司今天的發展”。
馬斯克回應稱,科斯拉說他后悔在2019年離開OpenAI根本是“不知所云”。
科斯拉的科斯拉風險投資公司(Khosla Ventures)是OpenAI的長期投資者。2019年,科斯拉風險投資公司在OpenAI投資5,000萬美元。同樣,他并不認同馬斯克的訴訟。科斯拉在X上發帖稱:“有人說,如果你沒有創新能力,那就去提起訴訟,這就是目前的狀況。”他在帖子中標記了馬斯克和OpenAI。
馬斯克加入之后,論戰仍在繼續。科斯拉依舊堅定地認為,人工智能比發明核彈更重要,因此不能完全開源,但他認同馬斯克和安德森的主張,認為領先的人工智能公司應該采取更嚴格的安全措施,甚至可以向政府求助。
科斯拉寫道:“我同意應該為所有[最先進的]人工智能提供國家網絡安全幫助和保護,而且這必不可少。人工智能不僅關乎網絡防御,還關乎誰將在全球經濟和政治競爭中勝出。全球價值觀和政治體制的未來都取決于人工智能。”
雖然科斯拉對于把所有人工智能研究開源持保留態度,但他表示不希望人工智能停止發展。他在回應安德森時表示:“[最先進的]人工智能不應該放慢開發速度,因為在我看來,敵對國家的危險程度更高。”
但在“人工智能對齊”這個問題上,科斯拉和安德森卻找到了一些共同點。“人工智能對齊”是指開發人工智能技術所使用的模型中的意識形態、原則和道德觀。(財富中文網)
譯者:劉進龍
審校:汪皓
維諾德·科斯拉與馬克·安德森都從公司創始人轉型為投資者。上周末,兩人就通用人工智能開發是否應該開源展開辯論。通用人工智能將使機器的智能水平與人類相當。
兩人爭論的導火索是科斯拉發布了一條帖子,對OpenAI及其首席執行官薩姆·奧爾特曼大加贊揚。
科斯拉寫道:“我們從@OpenAI創立之初就認識@sama,我們完全支持他和他的公司。這些訴訟是開發通用人工智能和實現其效益的巨大干擾。”
對于科斯拉的信息,安德森指責他“游說禁止開源”人工智能研究。
安德森似乎對科斯拉支持OpenAI的立場有不同意見,因為OpenAI已經背離了最初的開源原則。自人工智能誕生以來,安德森就堅定支持開源人工智能,他認為開源人工智能是一種保障措施,可以避免少數幾家大型科技公司和政府部門掌控最前沿人工智能研究的準入。
在當前的這場論戰和以前的言論中,安德森對于一些最大的人工智能批評者提出的擔憂嗤之以鼻。安德森曾經把這些擔憂歸咎于對顛覆性和不確定性的恐懼,而不是技術本身具有惡意,他在X上重申了這種觀點。
安德森在X上發帖稱:“每一種能夠提升人類福祉的重大新技術,都會引發虛假的道德恐慌。人工智能則是最新的例子。”
另一方面,科斯拉卻透過地緣政治和國家安全的視角看待人工智能,而不是從嚴格的創業角度來看待它。
對于安德森所說的科斯拉不支持開源的言論,科斯拉回應稱開源風險過高。
他回復安德森稱:“你會將曼哈頓計劃(Manhattan Project)開源嗎?人工智能對國家安全更加重要。我們必須在人工智能領域取得勝利。這是愛國主義,而不是口號。”
科斯拉和安德森之間的辯論,涉及到薩姆·奧爾特曼、OpenAI面臨的訴訟和埃隆·馬斯克。馬斯克后來也加入了論戰。兩人在辯論中還談到了是否應該允許任何人進行任何形式的人工智能研究,或者是否應該將最先進的人工智能版本交給政府。雖然這場辯論看起來只是一群非常成功的硅谷企業家在網絡上展開的一些討論,但我們可以管中窺豹,從中看到目前圍繞開源人工智能展開的關鍵辯論。
最終,沒有任何一方希望徹底禁止開源或閉源研究。在這場關于限制開源研究的論戰中,有部分主張源于擔心限制開源研究被作為一種惡意的理由,目的是確保對在人工智能領域已經取得進展的大公司實施監管。傳奇人工智能研究人員、Meta前首席人工智能科學家楊立昆(Yann LeCun)在X上加入論戰時表達了這種觀點。
他寫道:“沒有人要求禁止閉源人工智能。但有些人正在積極游說各國政府禁止(或限制)開源人工智能。有些人以軍事和經濟安全作為理由。還有人則提到了威脅人類生存的幻想。”
硅谷知名天使投資人羅恩·康威要求領先的人工智能公司承諾“開發能夠改善生活,為人類開啟更美好未來的人工智能”。迄今為止,Meta、谷歌(Google)、微軟(Microsoft)和OpenAI都已經在他的信中簽字。
安德森引用科斯拉使用的曼哈頓項目的比喻,對OpenAI的安全保障提出了擔憂。他認為,如果沒有與曼哈頓項目同等程度的安全保障,例如“嚴格的安全審查和許可流程”、“持續的內部監控”和有“全天候武裝警衛”守護的“牢不可破的物理設施”,那么OpenAI最先進的研究就將被美國的地緣政治對手們竊取。
OpenAI并未立即回復置評請求。
但安德森似乎不只是在爭論一種觀點,而是在進行思考練習。他在回應自己的帖子時寫道:“當然,每一個假設都很荒謬。”
埃隆·馬斯克加入論戰批評OpenAI的安全措施
此時,OpenAI的聯合創始人埃隆·馬斯克加入了論戰。
馬斯克回復安德森討論OpenAI安全措施的帖子稱:“國家行為者當然很容易盜取他們的知識產權。”
科斯拉也提到了馬斯克,稱馬斯克起訴OpenAI的決定是出于“酸葡萄心理”。上周,馬斯克起訴OpenAI,指控該初創公司違反了創業協議。馬斯克認為,OpenAI與微軟的密切關系和停止將其產品開源的決定,違背了公司創立的使命。據彭博社(Bloomberg)掌握的一份備忘錄顯示,OpenAI與科斯拉的觀點類似,指責馬斯克“后悔沒有參與公司今天的發展”。
馬斯克回應稱,科斯拉說他后悔在2019年離開OpenAI根本是“不知所云”。
科斯拉的科斯拉風險投資公司(Khosla Ventures)是OpenAI的長期投資者。2019年,科斯拉風險投資公司在OpenAI投資5,000萬美元。同樣,他并不認同馬斯克的訴訟。科斯拉在X上發帖稱:“有人說,如果你沒有創新能力,那就去提起訴訟,這就是目前的狀況。”他在帖子中標記了馬斯克和OpenAI。
馬斯克加入之后,論戰仍在繼續。科斯拉依舊堅定地認為,人工智能比發明核彈更重要,因此不能完全開源,但他認同馬斯克和安德森的主張,認為領先的人工智能公司應該采取更嚴格的安全措施,甚至可以向政府求助。
科斯拉寫道:“我同意應該為所有[最先進的]人工智能提供國家網絡安全幫助和保護,而且這必不可少。人工智能不僅關乎網絡防御,還關乎誰將在全球經濟和政治競爭中勝出。全球價值觀和政治體制的未來都取決于人工智能。”
雖然科斯拉對于把所有人工智能研究開源持保留態度,但他表示不希望人工智能停止發展。他在回應安德森時表示:“[最先進的]人工智能不應該放慢開發速度,因為在我看來,敵對國家的危險程度更高。”
但在“人工智能對齊”這個問題上,科斯拉和安德森卻找到了一些共同點。“人工智能對齊”是指開發人工智能技術所使用的模型中的意識形態、原則和道德觀。(財富中文網)
譯者:劉進龍
審校:汪皓
Vinod Khosla and Marc Andreessen, both founders turned investors, spent part of their weekends debating each other on whether the pursuit of artificial general intelligence—the idea that a machine could become as smart as a human—should be open-source.
The debate kicked off with a post from Khosla praising OpenAI and Sam Altman, the company’s CEO.
“We have known @sama since the early days of @OpenAI and fully support him and the company,” Khosla wrote. “These lawsuits are a massive distraction from the goals of getting to AGI and its benefits.”
Andreessen responded to Khosla’s message by accusing him of “lobbying to ban open source” research in AI.
Andreessen seemed to take issue with Khosla’s support for OpenAI because the firm has walked away from its previous open-source ethos. Since the advent of AI, Andreessen has come out as a big supporter of open-source AI, advocating it as a means to safeguard against a select few Big Tech firms and government agencies controlling access to the most cutting-edge AI research.
Both in this debate and in the past, Andreessen has been dismissive of the concerns raised by some of AI’s biggest critics. Andreessen has previously chalked up these worries to fears of disruption and uncertainty rather than the technology being malicious in and of itself—a point he reiterated in his exchange on X.
“Every significant new technology that advances human well-being is greeted by a ginned-up moral panic,” Andreessen posted on X. “This is just the latest.”
Khosla, on the other hand, tends to look at AI through a geopolitical and national-security lens rather than through a strictly entrepreneurial one.
In responding to Andreesen’s claims that he isn’t in favor of open-source, Khosla said the stakes were too high.
“Would you open source the Manhattan Project?” Khosla replied to Andreessen. “This one is more serious for national security. We are in a tech economic war with China and AI that is a must win. This is exactly what patriotism is about, not slogans.”
The back-and-forth discussion between Khosla and Andreessen saw the two opine on Sam Altman, OpenAI’s lawsuits, and Elon Musk, who chimed in himself at one point. The debate also explored whether anyone should be allowed to pursue any form of AI research, or if its most advanced versions should be delegated to the government. So while it may have seemed like just some online sniping between a group of extraordinarily successful Silicon Valley entrepreneurs, it contained a microcosm of the ongoing and critical debate around open-source AI.
Ultimately, neither camp wants to thoroughly ban open- or closed-source research. But part of the debate around limiting open-source research hinges on concerns it is being co-opted as a bad-faith argument to ensure regulatory capture for the biggest companies already making headway on AI—a point that legendary AI researcher and Meta’s former chief AI scientist Yann LeCun made when he entered the fray on X.
“No one is asking for closed-source AI to be banned,” LeCun wrote. “But some people are heavily lobbying governments around the world to ban (or limit) open source AI. Some of those people invoke military and economic security. Others invoke the fantasy of existential risk.”
Elsewhere in Silicon Valley, famed angel investor Ron Conway asked leading AI companies to pledge to “building AI that improves lives and unlocks a better future for humanity.” So far he has enlisted the likes of Meta, Google, Microsoft, and OpenAI as signatories to the letter.
Andreessen, sticking with Khosla’s Manhattan Project analogy, raised concerns about OpenAI’s safety protocols. He believes without the same level of security that surrounded the Manhattan Project—such as a “rigorous security vetting and clearance process,” “constant internal surveillance,” and “hardened physical facilities” with “24×7 armed guards”—OpenAI’s most advanced research could be stolen by the U.S.’s geopolitical rivals.
OpenAI did not immediately respond to a request for comment.
Andreessen, though, appears to have been doing more of a thought exercise than arguing a point, writing in response to his own post, “Of course every part of this is absurd.”
Elon Musk enters the debate to criticize OpenAI’s security
At this point, OpenAI cofounder Elon Musk chimed in.
“It would certainly be easy for a state actor to steal their IP,” Musk replied to Andreessen’s post about security at OpenAI.
Khosla, too, made mention of Musk, calling his decision to sue OpenAI “sour grapes.” Last week, Musk filed a lawsuit against OpenAI, alleging it breached the startup’s founding agreement. According to Musk, OpenAI’s close relationship with Microsoft and its decision to stop making its work open-source violated the organization’s mission. OpenAI took a similar tack to Khosla, accusing Musk of having “regrets about not being involved with the company today,” according to a memo obtained by Bloomberg.
Musk responded by saying Khosla “doesn’t know what he is talking about,” regarding his departure from OpenAI in 2019.
Khosla’s venture capital firm Khosla Ventures is a longtime backer of OpenAI. In 2019, Khosla Ventures invested $50 million into OpenAI. As such, he didn’t take kindly to Musk’s lawsuit. “Like they say if you can’t innovate, litigate and that’s what we have here,” Khosla wrote on X, tagging both Musk and OpenAI.
With Musk now involved, the debate continued. Khosla remained adamant AI was more important than the invention of the nuclear bomb and therefore couldn’t afford to be entirely open-source—though he did agree with Musk and Andreessen that its top firms should have more rigorous security measures, even relying on the government for assistance.
“Agree national cyber help and protection should be given and required for all [state of the art] AI,” Khosla wrote. “AI is not just cyber defense but also about winning economically and politically globally. The future of the world’s values and political system depends on it.”
Despite his reservations about making all of AI research open-source, Khosla said he did not want development to halt. “[State of the art] AI should not be slowed because enemy nation states are orders of magnitude bigger danger in my view,” Khosla said in response to Andreessen.
But Khosla and Andreessen did find some common ground on the question of AI alignment, which refers to the set of ideologies, principles, and ethics that will inform the models on which AI technologies are developed.