日韩中文字幕在线一区二区三区,亚洲热视频在线观看,久久精品午夜一区二区福利,精品一区二区三区在线观看l,麻花传媒剧电影,亚洲香蕉伊综合在人在线,免费av一区二区三区在线,亚洲成在线人视频观看
          首頁 500強(qiáng) 活動(dòng) 榜單 商業(yè) 科技 商潮 專題 品牌中心
          雜志訂閱

          科學(xué)家鉆研計(jì)算機(jī)的“大腦”,獲得驚人發(fā)現(xiàn)

          Jeremy Kahn
          2021-03-05

          人工智能研究公司OpenAI聲稱掌握了一種更先進(jìn)的方式,能夠用來一探“神經(jīng)網(wǎng)絡(luò)”這種人工智能軟件的內(nèi)部工作原理,從而有助于更好地了解“決策過程”這一著名的難題。

          文本設(shè)置
          小號(hào)
          默認(rèn)
          大號(hào)
          Plus(0條)

          研究人員稱,他們?nèi)〉昧艘豁?xiàng)重要的發(fā)現(xiàn)——可能對(duì)電腦、乃至人腦的研究都有著巨大深遠(yuǎn)的影響。

          位于舊金山的人工智能研究公司OpenAI表示,該公司掌握了一種更先進(jìn)的方式,能夠用來一探“神經(jīng)網(wǎng)絡(luò)”這種人工智能軟件的內(nèi)部工作原理,從而有助于他們更好地了解“決策過程”這一著名的難題。他們在研究中發(fā)現(xiàn),大型神經(jīng)網(wǎng)絡(luò)中的單個(gè)神經(jīng)元可以對(duì)某個(gè)特定的概念進(jìn)行編碼——這與神經(jīng)科學(xué)家在人腦中得到的發(fā)現(xiàn)驚人地相似。

          “神經(jīng)網(wǎng)絡(luò)”是一種對(duì)人腦的結(jié)構(gòu)進(jìn)行抽象、并以此建模而成的機(jī)器學(xué)習(xí)軟件。過去八年來,人工智能領(lǐng)域之所以能夠取得飛速的發(fā)展,主要就得益于神經(jīng)網(wǎng)絡(luò)的廣泛應(yīng)用——包括數(shù)字技術(shù)輔助的語音識(shí)別、面部識(shí)別軟件以及新的藥物研發(fā)方法等。

          但大型神經(jīng)網(wǎng)絡(luò)的一個(gè)缺點(diǎn)是,即便是對(duì)創(chuàng)造出它們的機(jī)器學(xué)習(xí)專家而言,理解其決策過程背后的原理也可能相當(dāng)困難。而這帶來的結(jié)果就是,很難確切知曉該軟件何時(shí)會(huì)發(fā)生故障,又會(huì)如何發(fā)生故障。因而也就不難理解,為什么這類人工智能明明可以比其他自動(dòng)化軟件或人工表現(xiàn)得更好,人們卻不太愿意使用它。這種情況在醫(yī)療和財(cái)務(wù)領(lǐng)域尤為普遍,因?yàn)橐坏Q策失誤,可能會(huì)付出金錢甚至生命的沉重代價(jià)。

          OpenAI的聯(lián)合創(chuàng)始人及首席科學(xué)家伊利亞·莎士科爾說:“由于我們不了解這些神經(jīng)網(wǎng)絡(luò)是如何工作的,因此當(dāng)它們報(bào)錯(cuò)時(shí),很難推理出是哪個(gè)環(huán)節(jié)出了問題。我們不知道它們是否可靠,或者是不是有一些在測試中看不出來的隱藏漏洞。”

          但最近,公司研究人員運(yùn)用多項(xiàng)技術(shù)來探究他們創(chuàng)建的一個(gè)大型神經(jīng)網(wǎng)絡(luò)的內(nèi)部工作原理,該神經(jīng)網(wǎng)絡(luò)是一個(gè)用來識(shí)別圖像并歸類存儲(chǔ)的軟件。他們的重要發(fā)現(xiàn)是,神經(jīng)網(wǎng)絡(luò)中的單個(gè)神經(jīng)元與某個(gè)特定的標(biāo)簽或概念相關(guān)。

          OpenAI在關(guān)于這項(xiàng)研究的一篇博文中說,這一點(diǎn)意義重大,因?yàn)樗?005年的一項(xiàng)具有里程碑意義的神經(jīng)科學(xué)研究發(fā)現(xiàn)如出一轍——2005年的研究是針對(duì)人腦展開的,其發(fā)現(xiàn)是,人腦中可能存在一種所謂的“祖母”神經(jīng)元,會(huì)對(duì)某個(gè)特定圖像或概念的刺激作出回應(yīng)。例如,神經(jīng)科學(xué)家發(fā)現(xiàn),其中一個(gè)研究對(duì)象似乎有一種與女明星哈莉·貝瑞有關(guān)的神經(jīng)元。當(dāng)向其展示哈莉·貝瑞的圖像時(shí),該神經(jīng)元就會(huì)被觸發(fā)、產(chǎn)生反應(yīng),當(dāng)這個(gè)人聽到“哈莉·貝瑞”一詞,或看到與哈莉·貝瑞有關(guān)的標(biāo)志性圖像時(shí),該神經(jīng)元也會(huì)被激活。

          OpenAI的研究對(duì)象是他們在今年1月首次推出的一項(xiàng)人工智能系統(tǒng)——該系統(tǒng)無需先對(duì)帶標(biāo)簽的數(shù)據(jù)集進(jìn)行分類識(shí)別訓(xùn)練,就能夠高度精確地執(zhí)行各種圖像歸類任務(wù)。這項(xiàng)被稱為CLIP(“語言-圖像比對(duì)預(yù)訓(xùn)練”的縮寫)的系統(tǒng)從網(wǎng)上攝取了4億張圖像,并與文案匹配。根據(jù)這些信息,該系統(tǒng)學(xué)會(huì)了從32768個(gè)文本摘要標(biāo)簽中預(yù)測,哪一個(gè)最可能與任意一個(gè)給定的圖像相關(guān)聯(lián)——即使此前從未見過這些圖像。例如,當(dāng)向CLIP展示一碗鱷梨醬的圖片時(shí),它不僅可以將其正確地標(biāo)記為鱷梨醬,而且還知道鱷梨醬是“一種食物”。

          OpenAI在新研究中使用的是一種“倒推”的方式,以探究是什么樣的圖片最能夠觸發(fā)某個(gè)人工神經(jīng)元的反應(yīng),讓機(jī)器認(rèn)為這是某個(gè)概念的“靈魂伴侶”。例如,OpenAI對(duì)與“金子”這一概念相關(guān)的神經(jīng)元進(jìn)行了探測,發(fā)現(xiàn)最可以觸發(fā)它的是含有閃亮的黃色硬幣狀物體和“金子”兩個(gè)字本身的圖片。而與“蜘蛛人”相關(guān)的神經(jīng)元?jiǎng)t會(huì)被一個(gè)漫畫人物的照片觸發(fā),也會(huì)被“蜘蛛”這個(gè)詞觸發(fā)。有趣的是,一個(gè)與“黃色”概念相關(guān)的神經(jīng)元會(huì)對(duì)“香蕉”和“檸檬”產(chǎn)生反應(yīng)——包括這兩個(gè)詞語本身和它們的顏色。

          OpenAI的研究人員加布里埃爾·高告訴《財(cái)富》雜志:“這或許證明了,這些神經(jīng)網(wǎng)絡(luò)并不像我們想象的那樣令人費(fèi)解。”將來,此類方法能夠幫助那些使用了神經(jīng)網(wǎng)絡(luò)技術(shù)的公司,讓他們了解這些機(jī)器是如何做出決策的,以及系統(tǒng)可能在何時(shí)出現(xiàn)故障或偏差。這還可以啟示神經(jīng)科學(xué)家們另辟蹊徑:用“人工”神經(jīng)網(wǎng)絡(luò)的原理,來研究“人類”的學(xué)習(xí)和概念形成方式。

          并非每個(gè)CLIP神經(jīng)元都只與一個(gè)特定的概念相關(guān)聯(lián)——很多會(huì)被不同概念激活并產(chǎn)生回應(yīng)。而且有些神經(jīng)元似乎能夠被同時(shí)觸發(fā),這或許意味著與之有關(guān)的是一個(gè)復(fù)雜的概念。

          OpenAI表示,也會(huì)有這樣一些概念,即研究人員原本期望系統(tǒng)中有與之相關(guān)的神經(jīng)元,但其實(shí)并不存在。例如盡管CLIP可以準(zhǔn)確地識(shí)別舊金山的照片,甚至通常能夠識(shí)別出拍攝照片的具體地點(diǎn),但神經(jīng)網(wǎng)絡(luò)似乎并沒有與“舊金山”甚至“城市”、“加利福尼亞”等概念本身相關(guān)的神經(jīng)元。OpenAI在其博文中說:“我們認(rèn)為,這些信息也被編碼在該模型的某個(gè)觸發(fā)系統(tǒng)中,但是以一種更獨(dú)特的方式編碼的。”

          這種技術(shù)還可以發(fā)現(xiàn)神經(jīng)網(wǎng)絡(luò)系統(tǒng)中潛藏的偏見。在演示中,研究人員發(fā)現(xiàn),CLIP還具有OpenAI所稱的一種“中東”神經(jīng)元。這種神經(jīng)元不僅會(huì)被與中東地區(qū)相關(guān)的圖像和單詞觸發(fā),并產(chǎn)生反應(yīng),也會(huì)和“恐怖分子”概念關(guān)聯(lián)。還有一個(gè)“移民”神經(jīng)元,與“拉丁美洲”關(guān)聯(lián)。還有能夠同時(shí)對(duì)深膚色人種和大猩猩產(chǎn)生反應(yīng)的神經(jīng)元——OpenAI指出,這種偏見和其他一些帶有種族主義色彩的圖片標(biāo)簽類似。此前,谷歌(Google)基于神經(jīng)網(wǎng)絡(luò)技術(shù)的圖像分類系統(tǒng)就出現(xiàn)過此類問題。

          大型人工智能系統(tǒng)中潛在的種族、性別偏見已經(jīng)引起了倫理學(xué)家和社會(huì)組織的日益關(guān)注,特別是那些通過互聯(lián)網(wǎng)大數(shù)據(jù)進(jìn)行機(jī)器學(xué)習(xí)、訓(xùn)練的人工智能系統(tǒng)。

          研究人員還說,他們的方法發(fā)現(xiàn)了CLIP在決策過程中存在的一項(xiàng)特殊偏見,這可能被有心人利用,故意誤導(dǎo)人工智能系統(tǒng)做出錯(cuò)誤的識(shí)別。該系統(tǒng)對(duì)表示某個(gè)概念的語詞或符號(hào)文本和那個(gè)概念本身緊密關(guān)聯(lián),以至于如果有人把這種符號(hào)或語詞放在不同的對(duì)象上,系統(tǒng)就會(huì)將其分錯(cuò)。例如,身上帶有巨大“$$$”標(biāo)志的狗可能就會(huì)被錯(cuò)誤地識(shí)別成“存錢罐”。

          加布里埃爾·高說:“我認(rèn)為,系統(tǒng)中肯定會(huì)帶有很多刻板印象。”而莎士科爾表示,能夠識(shí)別出這些偏差,正是試著糾正它們的第一步。他認(rèn)為,可以讓神經(jīng)網(wǎng)絡(luò)進(jìn)行一點(diǎn)附加的樣例訓(xùn)練,這些訓(xùn)練專門用來打破機(jī)器在學(xué)習(xí)中習(xí)得的不恰當(dāng)?shù)南嚓P(guān)性。(財(cái)富中文網(wǎng))

          編譯:陳聰聰

          研究人員稱,他們?nèi)〉昧艘豁?xiàng)重要的發(fā)現(xiàn)——可能對(duì)電腦、乃至人腦的研究都有著巨大深遠(yuǎn)的影響。

          位于舊金山的人工智能研究公司OpenAI表示,該公司掌握了一種更先進(jìn)的方式,能夠用來一探“神經(jīng)網(wǎng)絡(luò)”這種人工智能軟件的內(nèi)部工作原理,從而有助于他們更好地了解“決策過程”這一著名的難題。他們在研究中發(fā)現(xiàn),大型神經(jīng)網(wǎng)絡(luò)中的單個(gè)神經(jīng)元可以對(duì)某個(gè)特定的概念進(jìn)行編碼——這與神經(jīng)科學(xué)家在人腦中得到的發(fā)現(xiàn)驚人地相似。

          “神經(jīng)網(wǎng)絡(luò)”是一種對(duì)人腦的結(jié)構(gòu)進(jìn)行抽象、并以此建模而成的機(jī)器學(xué)習(xí)軟件。過去八年來,人工智能領(lǐng)域之所以能夠取得飛速的發(fā)展,主要就得益于神經(jīng)網(wǎng)絡(luò)的廣泛應(yīng)用——包括數(shù)字技術(shù)輔助的語音識(shí)別、面部識(shí)別軟件以及新的藥物研發(fā)方法等。

          但大型神經(jīng)網(wǎng)絡(luò)的一個(gè)缺點(diǎn)是,即便是對(duì)創(chuàng)造出它們的機(jī)器學(xué)習(xí)專家而言,理解其決策過程背后的原理也可能相當(dāng)困難。而這帶來的結(jié)果就是,很難確切知曉該軟件何時(shí)會(huì)發(fā)生故障,又會(huì)如何發(fā)生故障。因而也就不難理解,為什么這類人工智能明明可以比其他自動(dòng)化軟件或人工表現(xiàn)得更好,人們卻不太愿意使用它。這種情況在醫(yī)療和財(cái)務(wù)領(lǐng)域尤為普遍,因?yàn)橐坏Q策失誤,可能會(huì)付出金錢甚至生命的沉重代價(jià)。

          OpenAI的聯(lián)合創(chuàng)始人及首席科學(xué)家伊利亞·莎士科爾說:“由于我們不了解這些神經(jīng)網(wǎng)絡(luò)是如何工作的,因此當(dāng)它們報(bào)錯(cuò)時(shí),很難推理出是哪個(gè)環(huán)節(jié)出了問題。我們不知道它們是否可靠,或者是不是有一些在測試中看不出來的隱藏漏洞。”

          但最近,公司研究人員運(yùn)用多項(xiàng)技術(shù)來探究他們創(chuàng)建的一個(gè)大型神經(jīng)網(wǎng)絡(luò)的內(nèi)部工作原理,該神經(jīng)網(wǎng)絡(luò)是一個(gè)用來識(shí)別圖像并歸類存儲(chǔ)的軟件。他們的重要發(fā)現(xiàn)是,神經(jīng)網(wǎng)絡(luò)中的單個(gè)神經(jīng)元與某個(gè)特定的標(biāo)簽或概念相關(guān)。

          OpenAI在關(guān)于這項(xiàng)研究的一篇博文中說,這一點(diǎn)意義重大,因?yàn)樗?005年的一項(xiàng)具有里程碑意義的神經(jīng)科學(xué)研究發(fā)現(xiàn)如出一轍——2005年的研究是針對(duì)人腦展開的,其發(fā)現(xiàn)是,人腦中可能存在一種所謂的“祖母”神經(jīng)元,會(huì)對(duì)某個(gè)特定圖像或概念的刺激作出回應(yīng)。例如,神經(jīng)科學(xué)家發(fā)現(xiàn),其中一個(gè)研究對(duì)象似乎有一種與女明星哈莉·貝瑞有關(guān)的神經(jīng)元。當(dāng)向其展示哈莉·貝瑞的圖像時(shí),該神經(jīng)元就會(huì)被觸發(fā)、產(chǎn)生反應(yīng),當(dāng)這個(gè)人聽到“哈莉·貝瑞”一詞,或看到與哈莉·貝瑞有關(guān)的標(biāo)志性圖像時(shí),該神經(jīng)元也會(huì)被激活。

          OpenAI的研究對(duì)象是他們在今年1月首次推出的一項(xiàng)人工智能系統(tǒng)——該系統(tǒng)無需先對(duì)帶標(biāo)簽的數(shù)據(jù)集進(jìn)行分類識(shí)別訓(xùn)練,就能夠高度精確地執(zhí)行各種圖像歸類任務(wù)。這項(xiàng)被稱為CLIP(“語言-圖像比對(duì)預(yù)訓(xùn)練”的縮寫)的系統(tǒng)從網(wǎng)上攝取了4億張圖像,并與文案匹配。根據(jù)這些信息,該系統(tǒng)學(xué)會(huì)了從32768個(gè)文本摘要標(biāo)簽中預(yù)測,哪一個(gè)最可能與任意一個(gè)給定的圖像相關(guān)聯(lián)——即使此前從未見過這些圖像。例如,當(dāng)向CLIP展示一碗鱷梨醬的圖片時(shí),它不僅可以將其正確地標(biāo)記為鱷梨醬,而且還知道鱷梨醬是“一種食物”。

          OpenAI在新研究中使用的是一種“倒推”的方式,以探究是什么樣的圖片最能夠觸發(fā)某個(gè)人工神經(jīng)元的反應(yīng),讓機(jī)器認(rèn)為這是某個(gè)概念的“靈魂伴侶”。例如,OpenAI對(duì)與“金子”這一概念相關(guān)的神經(jīng)元進(jìn)行了探測,發(fā)現(xiàn)最可以觸發(fā)它的是含有閃亮的黃色硬幣狀物體和“金子”兩個(gè)字本身的圖片。而與“蜘蛛人”相關(guān)的神經(jīng)元?jiǎng)t會(huì)被一個(gè)漫畫人物的照片觸發(fā),也會(huì)被“蜘蛛”這個(gè)詞觸發(fā)。有趣的是,一個(gè)與“黃色”概念相關(guān)的神經(jīng)元會(huì)對(duì)“香蕉”和“檸檬”產(chǎn)生反應(yīng)——包括這兩個(gè)詞語本身和它們的顏色。

          OpenAI的研究人員加布里埃爾·高告訴《財(cái)富》雜志:“這或許證明了,這些神經(jīng)網(wǎng)絡(luò)并不像我們想象的那樣令人費(fèi)解。”將來,此類方法能夠幫助那些使用了神經(jīng)網(wǎng)絡(luò)技術(shù)的公司,讓他們了解這些機(jī)器是如何做出決策的,以及系統(tǒng)可能在何時(shí)出現(xiàn)故障或偏差。這還可以啟示神經(jīng)科學(xué)家們另辟蹊徑:用“人工”神經(jīng)網(wǎng)絡(luò)的原理,來研究“人類”的學(xué)習(xí)和概念形成方式。

          并非每個(gè)CLIP神經(jīng)元都只與一個(gè)特定的概念相關(guān)聯(lián)——很多會(huì)被不同概念激活并產(chǎn)生回應(yīng)。而且有些神經(jīng)元似乎能夠被同時(shí)觸發(fā),這或許意味著與之有關(guān)的是一個(gè)復(fù)雜的概念。

          OpenAI表示,也會(huì)有這樣一些概念,即研究人員原本期望系統(tǒng)中有與之相關(guān)的神經(jīng)元,但其實(shí)并不存在。例如盡管CLIP可以準(zhǔn)確地識(shí)別舊金山的照片,甚至通常能夠識(shí)別出拍攝照片的具體地點(diǎn),但神經(jīng)網(wǎng)絡(luò)似乎并沒有與“舊金山”甚至“城市”、“加利福尼亞”等概念本身相關(guān)的神經(jīng)元。OpenAI在其博文中說:“我們認(rèn)為,這些信息也被編碼在該模型的某個(gè)觸發(fā)系統(tǒng)中,但是以一種更獨(dú)特的方式編碼的。”

          這種技術(shù)還可以發(fā)現(xiàn)神經(jīng)網(wǎng)絡(luò)系統(tǒng)中潛藏的偏見。在演示中,研究人員發(fā)現(xiàn),CLIP還具有OpenAI所稱的一種“中東”神經(jīng)元。這種神經(jīng)元不僅會(huì)被與中東地區(qū)相關(guān)的圖像和單詞觸發(fā),并產(chǎn)生反應(yīng),也會(huì)和“恐怖分子”概念關(guān)聯(lián)。還有一個(gè)“移民”神經(jīng)元,與“拉丁美洲”關(guān)聯(lián)。還有能夠同時(shí)對(duì)深膚色人種和大猩猩產(chǎn)生反應(yīng)的神經(jīng)元——OpenAI指出,這種偏見和其他一些帶有種族主義色彩的圖片標(biāo)簽類似。此前,谷歌(Google)基于神經(jīng)網(wǎng)絡(luò)技術(shù)的圖像分類系統(tǒng)就出現(xiàn)過此類問題。

          大型人工智能系統(tǒng)中潛在的種族、性別偏見已經(jīng)引起了倫理學(xué)家和社會(huì)組織的日益關(guān)注,特別是那些通過互聯(lián)網(wǎng)大數(shù)據(jù)進(jìn)行機(jī)器學(xué)習(xí)、訓(xùn)練的人工智能系統(tǒng)。

          研究人員還說,他們的方法發(fā)現(xiàn)了CLIP在決策過程中存在的一項(xiàng)特殊偏見,這可能被有心人利用,故意誤導(dǎo)人工智能系統(tǒng)做出錯(cuò)誤的識(shí)別。該系統(tǒng)對(duì)表示某個(gè)概念的語詞或符號(hào)文本和那個(gè)概念本身緊密關(guān)聯(lián),以至于如果有人把這種符號(hào)或語詞放在不同的對(duì)象上,系統(tǒng)就會(huì)將其分錯(cuò)。例如,身上帶有巨大“$$$”標(biāo)志的狗可能就會(huì)被錯(cuò)誤地識(shí)別成“存錢罐”。

          加布里埃爾·高說:“我認(rèn)為,系統(tǒng)中肯定會(huì)帶有很多刻板印象。”而莎士科爾表示,能夠識(shí)別出這些偏差,正是試著糾正它們的第一步。他認(rèn)為,可以讓神經(jīng)網(wǎng)絡(luò)進(jìn)行一點(diǎn)附加的樣例訓(xùn)練,這些訓(xùn)練專門用來打破機(jī)器在學(xué)習(xí)中習(xí)得的不恰當(dāng)?shù)南嚓P(guān)性。(財(cái)富中文網(wǎng))

          編譯:陳聰聰

          Researchers say they have made an important finding that could have big implications for the study of computer brains and, possibly, human ones too.

          OpenAI, the San Francisco–based A.I. research company, says that it has advanced methods for peering into the inner workings of artificial intelligence software known as neural networks, helping to make their notoriously opaque decision-making more interpretable. In the process they have uncovered that individual neurons in a large neural network can encode a particular concept, a finding that parallels one that neuroscientists have glimpsed in the human brain.

          Neural networks are a kind of machine-learning software loosely modeled on the human brain. The use of these networks has been responsible for most of the rapid advances in artificial intelligence in the past eight years, including the speech recognition found in digital assistants, facial recognition software, and new ways to discover drugs.

          But one drawback in large neural networks is that it can be challenging to understand the rationale behind their decisions, even for the machine-learning experts who create them. As a result, it is difficult to know exactly when and how this software can fail. And that has made people understandably reluctant to use such A.I. software, even when these A.I. systems seem to outperform other kinds of automated software or humans. This has particularly been true in medical and financial settings, where a wrong decision may cost money or even lives.

          “Because we don’t understand how these neural networks work, it can be hard to reason about their errors,” says Ilya Sutskever, OpenAI’s cofounder and chief scientist. “We don’t know if they are reliable or if they have hidden vulnerabilities that are not apparent from testing.”

          But researchers at the company recently used several techniques to probe the inner workings of a large neural network they had created for identifying images and putting them into broad category buckets. The researchers discovered that individual neurons in the network were associated with one particular label or concept.

          This was significant, OpenAI said in a blog post discussing its research, because it echoed findings from a landmark 2005 neuroscience study that found the human brain may have “grandmother” neurons that fire in response to one very specific image or concept. For instance, the neuroscientists discovered that one subject in their study seemed to have a neuron that was associated with the actress Halle Berry. The neuron fired when the person was shown an image of Berry, but the same neuron was also activated when the person heard the words “Halle Berry,” or when shown images associated with Berry’s iconic roles.

          OpenAI’s research focused on an A.I. system it debuted in January that can perform a wide variety of image classification tasks with a high degree of accuracy, without being specifically trained for those tasks with labeled data sets. The system, called CLIP (short for Contrastive Language-Image Pre-training), ingested 400 million images from the Internet and paired with captions. From this information, the technology learned to predict which of 32,768 text snippet labels was most likely to be associated with any given image, even those it had never encountered before. For instance, show CLIP a picture of a bowl of guacamole, and not only is it able to correctly label the image as guacamole but it also knows that guacamole is “a type of food.”

          In the new research, OpenAI used techniques that reverse engineer what makes a particular artificial neuron fire the most to build up a picture of that neuron’s “Platonic ideal” for a given concept. For instance, OpenAI probed one neuron associated with the concept “gold” and found that the image that most activated it would contain shiny yellow coin-like objects as well as a picture of the text “gold” itself. A neuron affiliated with “spider man” was triggered in response to photos of a person dressed up as the comic book hero but also to the word “spider.” Interestingly, a neuron affiliated with the concept “yellow” fired in response to the words “banana” and “l(fā)emon,” as well as the color and word itself.

          “This is maybe evidence that these neural networks are not as incomprehensible as we might think,” Gabriel Goh, the OpenAI researcher who led the team working on interpreting CLIP’s conceptual reasoning, told Fortune. In the future, such methods could be used to help companies using neural networks to understand how they arrive at decisions and when a system is likely to fail or exhibit bias. It might also point a way for neuroscientists to use artificial neural networks to investigate the ways in which human learning and concept formation may take place.

          Not every CLIP neuron was associated with a distinct concept. Many fired in response to a number of different conceptual categories. And some neurons seemed to fire together, possibly meaning that they represented a complex concept.

          OpenAI said that some concepts that the researchers expected the system to have a neuron for were absent. Even though CLIP can accurately identify photographs of San Francisco and can often even identify the neighborhood of the city in which they were taken, the neural network did not seem to have a neuron associated with a concept of “San Francisco” or even “city” or “California.” “We believe this information to be encoded within the activations of the model somewhere, but in a more exotic way,” OpenAI said in its blog post.

          In a demonstration that this technique can be used to uncover hidden biases in neural networks, the researchers discovered that CLIP also had what OpenAI dubbed a “Middle East” neuron that fired in response to images and words associated with the region, but also in response to those associated with terrorism, the company said. It had an “immigration” neuron that responds to Latin America. And it found a neuron that fired for both dark-skinned people and gorillas, which OpenAI noted was similar to other racist photo tagging that had previously caused problems for neural network–based image classification systems at Google.

          Racial and gender biases hidden in large A.I. models have become an increasing area of concern for A.I. ethics researchers and civil society organizations, especially those trained from massive amount of data culled from the Internet.

          The researchers also said that their methods had uncovered a particular bias in how CLIP makes decisions that would make it possible for someone to fool the A.I. into making incorrect identifications. The system associated the text of a word or symbol associated with a concept so strongly that if someone put that symbol or word on a different object, the system would misclassify it. For instance, a dog with a big “$$$” sign on it might be misclassified as a piggy bank.

          “I think you definitely see a lot of stereotyping in the model,” Goh said. Sutskever said that being able to identify these biases was a first step toward trying to correct them, something he thought could be accomplished by providing the neural network with a relatively small number of additional training examples that are specifically designed to break the inappropriate correlation the system has learned.

          財(cái)富中文網(wǎng)所刊載內(nèi)容之知識(shí)產(chǎn)權(quán)為財(cái)富媒體知識(shí)產(chǎn)權(quán)有限公司及/或相關(guān)權(quán)利人專屬所有或持有。未經(jīng)許可,禁止進(jìn)行轉(zhuǎn)載、摘編、復(fù)制及建立鏡像等任何使用。
          0條Plus
          精彩評(píng)論
          評(píng)論

          撰寫或查看更多評(píng)論

          請(qǐng)打開財(cái)富Plus APP

          前往打開