




版權(quán)說(shuō)明:本文檔由用戶提供并上傳,收益歸屬內(nèi)容提供方,若內(nèi)容存在侵權(quán),請(qǐng)進(jìn)行舉報(bào)或認(rèn)領(lǐng)
文檔簡(jiǎn)介
TheEUandU.S.divergeonAI
regulation:Atransatlanticcomparisonandstepstoalignment
EXECUTIVESUMMARY
TheEUandtheU.S.arejointlypivotaltothefutureofglobalAIgovernance.EnsuringthatEUandU.S.approachestoAIriskmanagementaregenerallyalignedwillfacilitatebilateraltrade,improveregulatoryoversight,andenablebroadertransatlanticcooperation.
TheU.S.approachtoAIriskmanagementishighlydistributedacrossfederalagencies,manyadaptingtoAIwithoutnewlegalauthorities.Meanwhile,theU.S.hasinvestedinnon-regulatoryinfrastructure,suchasanewAIriskmanagementframework,evaluationsoffacialrecognitionsoftware,andextensivefundingofAIresearch.TheEUapproachtoAIriskmanagementischaracterizedbyamorecomprehensiverangeoflegislationtailoredtospecificdigitalenvironments.TheEUplanstoplacenewrequirementsonhigh-riskAIinsocioeconomicprocesses,thegovernmentuseofAI,andregulatedconsumerproductswithAIsystems.OtherEUlegislationenablesmorepublictransparencyandinfluenceoverthedesignofAIsystemsinsocialmediaande-commerce.
TheEUandU.S.strategiesshareaconceptualalignmentonarisk-basedapproach,agreeonkeyprinciplesoftrustworthyAI,andendorseanimportantroleforinternationalstandards.However,thespecificsoftheseAIriskmanagementregimeshavemoredifferencesthansimilarities.RegardingmanyspecificAIapplications,especiallythoserelatedtosocioeconomicprocessesandonlineplatforms,theEUandU.S.areonapathtosignificantmisalignment.
TheEU-U.S.TradeandTechnologyCouncilhasdemonstratedearlysuccessworkingonAI,especiallyonaprojecttodevelopacommonunderstandingofmetricsandmethodologiesfortrustworthyAI.Throughthesenegotiations,theEUandU.S.havealsoagreedtoworkcollaborativelyoninternationalAIstandards,whilealsojointlystudyingemergingrisksofAIandapplicationsofnewAItechnologies.
MorecanbedonetofurthertheEU-U.S.alignment,whilealsoimprovingeachcountry’sAIgovernanceregime.Specifically:
oTheU.S.shouldexecuteonfederalagencyAIregulatoryplansanduse
thesefordesigningstrategicAIgovernancewithaneyetowardsEU-U.S.alignment.
oTheEUshouldcreatemoreflexibilityinthesectoralimplementationof
theEUAIAct,improvingthelawandenablingfutureEU-U.S.cooperation.
oTheU.S.needstoimplementalegalframeworkforonlineplatform
governance,butuntilthen,theEUandU.S.shouldworkonshareddocumentationofrecommendersystemsandnetworkalgorithms,aswellasperformcollaborativeresearchononlineplatforms.
oTheU.S.andEUshoulddeepenknowledgesharingonanumberoflevels,
includingonstandardsdevelopment;AIsandboxes;largepublicAIresearchprojectsandopen-sourcetools;regulator-to-regulatorexchanges;anddevelopinganAIassuranceecosystem.
MorecollaborationbetweentheEUandtheU.S.willbecrucial,asthesegovernmentsareimplementingpoliciesthatwillbefoundationaltothe
democraticgovernanceofAI.
INTRODUCTION
Approachestoartificialintelligence(AI)riskmanagement—shapedbyemerginglegislation,regulatoryoversight,civilliability,softlaw,andindustrystandards—arebecomingkeyfacetsofinternationaldiplomacyandtradepolicy.Inadditiontoencouragingintegratedtechnologymarkets,amoreunifiedinternationalapproachtoAIgovernancecanstrengthenregulatoryoversight,guideresearchtowardssharedchallenges,promotetheexchangeofbestpractices,andenabletheinteroperabilityoftoolsfortrustworthyAIdevelopment.
EspeciallyimpactfulinthislandscapearetheEUandtheU.S.,whicharebothcurrentlyimplementingfoundationalpoliciesthatwillsetprecedentsforthefutureofAIriskmanagementwithintheirterritoriesandglobally.ThegovernanceapproachesoftheEUandU.S.touchonawiderangeofAIapplicationswithinternationalimplications,includingmoresophisticatedAIinconsumerproducts;aproliferationofAIinregulatedsocioeconomicdecisions;anexpansionofAIinawidevariety
ofonlineplatforms;andpublic-facingweb-hostedAIsystems,suchas
generativeAIandfoundationmodels
.[i]
Thispaperconsidersthebroad
approachesoftheU.S.andtheEUtoAIriskmanagement,comparespolicydevelopmentsacrosseightkeysubfields,anddiscussescollaborativestepstakensofar,especiallythroughtheEU-U.S.TradeandTechnologyCouncil.Further,thispaperidentifieskeyemergingchallengestotransatlanticAIriskmanagementandofferspolicymakingrecommendationsthatmightadvancewell-alignedandmutually
beneficialEU-U.S.AIpolicy.
THEU.S.APPROACHTOAIRISKMANAGEMENT
TheU.S.federalgovernment’sapproachtoAIriskmanagementcanbroadlybecharacterizedasrisk-based,sectorallyspecific,andhighlydistributedacrossfederalagencies.Thereareadvantagestothisapproach,howeveritalsocontributestotheunevendevelopmentofAI
policies.WhilethereareseveralguidingfederaldocumentsfromtheWhiteHouseonAIharms,theyhavenotcreatedanevenorconsistent
federalapproachtoAIrisks.
“Byandlarge,federalagencieshavestillnotdevelopedtherequiredAIregulatoryplans.”
TheFebruary2019executiveorder,MaintainingAmericanLeadershipinArtificialIntelligence(EO13859),anditsensuingOfficeofManagement
andBudget(OMB)guidance(M-21-06)presentedthefirstfederal
approachtoAIoversight
.[1]
DeliveredinNovember2020,15
monthsafterthedeadlinesetinEO13859,theOMBguidanceclearlyarticulatedarisk-basedapproach,stating“themagnitudeandnatureoftheconsequencesshouldanAItoolfail…canhelpinformthelevelandtypeofregulatoryeffortthatisappropriatetoidentifyandmitigaterisks.”ThesedocumentsalsourgedagenciestoconsiderkeyfacetsofAIriskreductionthroughregulatoryandnon-regulatoryinterventions.ThisincludesusingscientificevidencetodetermineAI’scapabilities,enforcingnon-discriminationstatutes,consideringdisclosurerequirements,andpromotingsafeAIdevelopmentanddeployment.WhilethesedocumentsreflectedtheTrumpadministration’sminimalist
regulatoryperspective,theyalsorequiredagenciestodevelopplansto
regulateAIapplications
.[2]
Byandlarge,federalagencieshavestillnotdevelopedtherequiredAIregulatoryplans.InDecember2022,StanfordUniversity’sCenterfor
Human-CenteredAIreleasedareportstatingthatonlyfiveof41major
agenciescreatedanAIplanasrequired
.[3,]
[ii]
Thisisagenerous
interpretation,asonlyonemajoragency,theDepartmentofHealthand
HumanServices(HHS),providedathoroughplaninresponse
.[4]
HHS
extensivelydocumentedtheagency’sauthorityoverAIsystems(through12differentstatutes),itsactiveinformationcollections(e.g.,onAIforgenomicsequencing),andtheemergingAIusecasesofinterest(mostlyinillnessdetection).ThethoroughnessoftheHHS’sregulatoryplanshowshowvaluablethisendeavorcouldbeforfederalagencyplanningandinformingthepublicifotheragenciesweretofollowinHHS’sfootsteps.
RatherthanfurtherimplementingEO13859,theBidenadministration
insteadrevisitedthetopicofAIrisksthroughtheBlueprintforanAIBill
ofRights(AIBoR)
.[5]
DevelopedbytheWhiteHouseOfficeofScienceand
TechnologyPolicy(OSTP),theAIBoRincludesadetailedexpositionofAIharmstoeconomicandcivilrights,fiveprinciplesformitigatingtheseharms,andanassociatedlistoffederalagencies’actions.TheAIBoR
endorsesasectorallyspecificapproachtoAIgovernance,withpolicyinterventionstailoredtoindividualsectorssuchashealth,labor,andeducation.Itsapproachisthereforequitereliantontheseassociatedfederalagencyactions,ratherthancentralizedaction,especiallybecausetheAIBoRisnonbindingguidance.
ThattheAIBoRdoesnotdirectlycompelfederalagenciestomitigateAI
risksisclearfromthepatchworkofresponses,withsignificanteffortsin
someagenciesandnon-responseinothers
.[6]
Further,despitethefivebroadprinciplesoutlinedintheAIBoR
,[iii]
mostfederalagenciesare
onlyabletoadapttheirpre-existinglegalauthoritiestoalgorithmicsystems.ThisisbestdemonstratedbyagenciesregulatingAIusedtomakesocioeconomicdecisions.ThisincludestheFederalTradeCommission(FTC),whichcanuseitsauthoritytoprotectagainst“unfair
anddeceptive”practicestoenforcetruthinadvertisingandsomedata
privacyguaranteesinAIsystems
.[7]
TheFTCisalsoactivelyconsidering
howitsexistingauthoritiesaffectdata-drivencommercialsurveillance,includingalgorithmicdecision-making,andsomeadvocacy
organizationshavearguedtheFTCcanplacetransparencyandfairness
requirementsonsuchalgorithmicsystems
.[8]
TheEqualEmployment
OpportunityCommission(EEOC)canimposesometransparency,requireanon-AIalternativeforpeoplewithdisabilities,andenforcenon-
discriminationinAIhiring
.[9]
TheConsumerFinancialProtectionBureau
(CFPB)requiresexplanationsforcreditdenialsfromAIsystemsand
couldpotentiallyenforcenon-discriminationrequirements
.[10]
Thereare
otherexamples,however,innosectordoesanyagencyhavethelegalauthoritiesnecessarytoenforcealloftheprinciplesexpressedbytheAIBoR,northoseinEO13859.
Oftheseprinciples,theBidenadministrationhasbeenespeciallyvocalonracialequityandinFebruary2023publishedtheexecutiveorderFurtherAdvancingRacialEquityandSupportforUnderservedCommunitiesThroughtheFederalGovernment(EO14091).Thesecondexecutiveorderonthissubject,EO14091,directsfederalagenciesto
addressemergingriskstocivilrights,including“algorithmic
discriminationinautomatedtechnology.
”[11]
Itistoosoontoknowthe
impactofthisnewexecutiveorder.
Federalagencieswithregulatorypurviewoverconsumerproductsarealsomakingadjustments.OneleadingagencyistheFoodandDrugAdministration(FDA),whichhasbeenworkingtoincorporateAI,and
specificallymachinelearning,inmedicaldevicessinceatleast
2019
.[12]
TheFDAnowpublishesbestpracticesforAIinmedicaldevices,
documentscommerciallyavailableAI-enabledmedicaldevices,andhaspromisedtoperformrelevantpilotsandadvanceregulatorysciencein
itsAIactionplan
.[13]
AsidefromtheFDA,theConsumerProductsSafety
Commission(CPSC)statedin2019itsintentiontoresearchandtrackincidentsofAIharmsinconsumerproducts,aswellastoconsiderpolicy
interventionsincludingpubliceducationcampaigns,voluntary
standards,mandatorystandards,andpursuingrecalls
.[14]
In2022,CPSC
issuedadraftreportonhowtotestandevaluateconsumerproducts
whichincorporatemachinelearning
.[15]
Issuedinthefinaldaysofthe
Trumpadministration,theDepartmentofTransportation’sAutomated
VehiclesComprehensivePlansoughttoremoveregulatoryrequirements
forsemi-andfully-autonomousvehicles
.[16]
InparallelwiththeunevenstateofAIregulatorydevelopments,theU.S.iscontinuingtoinvestininfrastructureformitigatingAIrisks.MostnotableistheNationalInstituteofStandardsandTechnology’s(NIST)AI
RiskManagementFramework(RMF),firstreleasedasadraftonMarch
17,2022,withafinalreleaseonJanuary26,2023
.[17]
TheNISTAIRMFis
avoluntaryframeworkthatbuildsofftheOrganizationforEconomicCooperationandDevelopment’s(OECD)Frameworkforthe
ClassificationofAISystemsbyofferingcomprehensivesuggestionson
whenandhowriskcanbemanagedthroughouttheAIlifecycle
.[18]
NIST
isalsodevelopinganewAIRMFPlaybook,withconcreteexamplesofhowentitiescanimplementtheRMFacrossthedatacollection,
development,deployment,andoperationofAI
.[19]
TheNISTAIRMFwill
alsobeaccompaniedbyaseriesofcasestudies,eachofwhichwill
documentthestepsandinterventionstakentomitigateriskwithina
specificAIapplication
.[20]
Whileitistoosoontotellwhatdegreeof
adoptiontheNISTAIRMFwillachieve,the2014NISTCybersecurity
Frameworkhasbeenwidelyadapted(usuallyentailingpartialadoption)
byindustry
.[21]
NISTalsoplaysaroleinevaluatingandpubliclyreportingonthe
accuracyandfairnessoffacialrecognitionalgorithmsthroughits
ongoingFaceRecognitionVendorTestprogram
.[22]
Inoneanalysis,NIST
testedandcompared189commercialfacialrecognitionalgorithmsforaccuracyondifferentdemographicgroups,contributingvaluable
informationtotheAImarketplaceandimprovingpublicunderstanding
ofthesetools
.[23]
Anassortmentofotherpolicyactionsaddressessomealgorithmicharmsandcontributestofutureinstitutionalpreparednessandthuswarrantsmention,evenifAIriskisnottheprimaryorientation.LaunchedinApril2022,theNationalAIAdvisoryCommitteemayplayanexternaladvisoryroleinguidinggovernmentpolicyonmanagingAIrisksinareassuchas
lawenforcement,althoughitisprimarilyconcernedwithadvancingAIas
anationaleconomicresource
.[24]
Thefederalgovernmenthasalsorun
severalpilotsofanimprovedhiringprocess,aimedatattractingdata
sciencetalenttothecivilservice,akeyaspectofpreparednessforAI
governance
.[25]
Currently,the“datascientist”occupationalseriesisthe
mostrelevantfederalgovernmentjobforthetechnicalaspectsofAIriskmanagement.However,thisroleismoreorientedtowardsperforming
datasciencethanreviewingorauditingAImodelscreatedbyprivate
sectordatascientists
.[26]
[iv]
TheU.S.governmentfirstpublishedanationalAIResearchand
DevelopmentStrategicPlanin2016,andin2022,13federaldepartments
fundedAIresearchanddevelopment
.[27]
TheNationalScience
Foundationhasnowfunded19interdisciplinaryAIresearchinstitutes,
andtheacademicworkcomingfromsomeoftheseinstitutesis
advancingtrustworthyandethicalAImethods
.[28]
Similarly,the
DepartmentofEnergywastaskedwithdevelopingmorereliableAI
methodswhichmightinformcommercialactivity,suchasinmaterials
discovery
.[29]
Further,theBidenadministrationwillseekanadditional
$2.6billionoversixyearstofundAIinfrastructureundertheNationalAI
ResearchResource(NAIRR)project,whichstatesthatencouraging
trustworthyAIisoneofitsfourkeygoals
.[30]
Specifically,theNAIRR
couldbeusedtobetterstudytherisksofemerginglargeAImodels,manyofwhicharecurrentlydevelopedwithoutpublicscrutiny.
Inasignificantrecentdevelopment,aseriesofstateshaveintroduced
legislationtotacklealgorithmicharms,includingCalifornia,Connecticut,
andVermont
.[31]
WhilethesemightmeaningfullyimproveAIprotections,
theycouldalsopotentiallyleadtofuturepre-emptionissuesthatwouldmirrortheongoingchallengetopassingfederalprivacylegislation
(namely,howshouldthefederallegislationreplaceoraugmentvarious
statelaws)
.[32]
THEEUAPPROACHTOAIRISKMANAGEMENT
TheEU’sapproachtoAIriskmanagementiscomplexandmultifaceted,buildingonimplementedlegislation,especiallytheGeneralDataProtectionRegulation(GDPR),andspanningnewlyenactedlegislation,namelytheDigitalServicesActandDigitalMarketsAct,aswellaslegislationstillbeingactivelydebated,particularlytheAIAct,amongotherrelevantendeavors.TheEUhasconsciouslydevelopeddifferentregulatoryapproachesfordifferentdigitalenvironments,eachwitha
differentdegreeofemphasisonAI.
“TheEUhasconsciouslydevelopeddifferentregulatoryapproachesfordifferentdigitalenvironments,eachwithadifferentdegreeofemphasisonAI.”
Asidefromitsdataprivacyimplications,GPDRcontainstwoimportantarticlesrelatedtoalgorithmicdecision-making.First,GDPRstatesthat
algorithmicsystemsshouldnotbeallowedtomakesignificantdecisions
thataffectlegalrightswithoutanyhumansupervision
.[33]
Basedonthis
clause,in2021,Uberwasrequiredtoreinstatesixdriverswhowere
foundtohavebeenfiredsolelybythecompany’salgorithmic
system
.[34]
Second,GDPRguaranteesanindividual’srightto“meaningful
informationaboutthelogic”ofalgorithmicsystems,attimes
controversiallydeemeda“righttoexplanation.
”[35]
Inpractice,
companiessuchashomeinsuranceprovidershaveofferedlimited
responsestorequestsforinformationaboutalgorithmic
decisions
.[36]
Therearemanyopenquestionsaboutthisclause,including
howoftenaffectedindividualsrequestthisinformation,howvaluablethe
informationistothem,andwhathappenswhencompaniesrefuseto
providetheinformation
.[37]
TheEUAIActwillbeanespeciallycriticalcomponentoftheEU’s
approachtoAIriskmanagementinmanyareasofAIrisk
.[38]
WhiletheAI
Actisnotyetfinalized,enoughcanbeinferredfromtheEuropeanCommissionproposalfromApril2021,thefinalCounciloftheEUproposalfromDecember2022,andtheavailableinformationfromtheongoingEuropeanParliamentdiscussionsinordertoanalyzeitskeyfeatures.
Althoughitisoftenreferredtoas“horizontal,”theAIActimplementsa
tieredsystemofregulatoryobligationsforaspecificallyenumeratedlist
ofAIapplications
.[39]
SeveralAIapplications,includingdeepfakes,
chatbots,andbiometricanalysis,mustclearlydisclosethemselvestoaffectedpersons.AdifferentsetofAIsystemswith“unacceptablerisks”
wouldbebannedcompletely,potentiallyincludingAIforsocial
scoring
,[v]
AI-enabledmanipulativetechnologies,and,withseveral
importantexceptions,biometricidentificationbylawenforcementinpublicspaces.
Betweenthesetwotierssits“high-risk”AIsystems,whichisthemostinclusiveandimpactfulofthedesignationsintheEUAIAct.TwocategoriesofAIapplicationswillbedesignatedashigh-riskundertheAIAct:regulatedconsumerproductsandAIusedforimpactfulsocioeconomicdecisions.Allhigh-riskAIsystemswillhavetomeetstandardsofdataquality,accuracy,robustness,andnon-discrimination,whilealsoimplementingtechnicaldocumentation,record-keeping,ariskmanagementsystem,andhumanoversight.Entitiesthatsellordeploycoveredhigh-riskAIsystems,calledproviders,willneedtomeettheserequirementsandsubmitdocumentationthatattesttotheconformityoftheirAIsystemsorotherwisefacefinesashighas6%ofannualglobalturnover.
Thefirstcategoryofhigh-riskAIincludesconsumerproductsthatarealreadyregulatedundertheNewLegislativeFramework,theEU’ssingle-
marketregulatoryregime,whichincludesproductssuchasmedical
devices,vehicles,boats,toys,andelevators
.[40]
Generallyspeaking,this
meansthatAI-enabledconsumerproductswillstillgothroughthepre-existingregulatoryprocessunderthepertinentproductharmonizationlegislationandwillnotneedasecond,independentconformityassessmentjustfortheAIActrequirements.Therequirementsforhigh-riskAIsystemswillbeincorporatedintotheexistingproductharmonizationlegislation.Asaresult,ingoingthroughthepre-existingregulatoryprocess,businesseswillhavetopaymoreattentiontoAIsystems,reflectingthefactthatsomemodernAIsystemsmaybemoreopaque,lesspredictable,orplausiblyupdateafterthepointofsale.Notably,someEUagencieshavealreadybeguntoconsiderhowAIaffectstheirregulatoryprocesses.Oneleadingex
溫馨提示
- 1. 本站所有資源如無(wú)特殊說(shuō)明,都需要本地電腦安裝OFFICE2007和PDF閱讀器。圖紙軟件為CAD,CAXA,PROE,UG,SolidWorks等.壓縮文件請(qǐng)下載最新的WinRAR軟件解壓。
- 2. 本站的文檔不包含任何第三方提供的附件圖紙等,如果需要附件,請(qǐng)聯(lián)系上傳者。文件的所有權(quán)益歸上傳用戶所有。
- 3. 本站RAR壓縮包中若帶圖紙,網(wǎng)頁(yè)內(nèi)容里面會(huì)有圖紙預(yù)覽,若沒(méi)有圖紙預(yù)覽就沒(méi)有圖紙。
- 4. 未經(jīng)權(quán)益所有人同意不得將文件中的內(nèi)容挪作商業(yè)或盈利用途。
- 5. 人人文庫(kù)網(wǎng)僅提供信息存儲(chǔ)空間,僅對(duì)用戶上傳內(nèi)容的表現(xiàn)方式做保護(hù)處理,對(duì)用戶上傳分享的文檔內(nèi)容本身不做任何修改或編輯,并不能對(duì)任何下載內(nèi)容負(fù)責(zé)。
- 6. 下載文件中如有侵權(quán)或不適當(dāng)內(nèi)容,請(qǐng)與我們聯(lián)系,我們立即糾正。
- 7. 本站不保證下載資源的準(zhǔn)確性、安全性和完整性, 同時(shí)也不承擔(dān)用戶因使用這些下載資源對(duì)自己和他人造成任何形式的傷害或損失。
最新文檔
- 二零二五年度文化娛樂(lè)產(chǎn)業(yè)股權(quán)轉(zhuǎn)讓及代持合作協(xié)議
- 2025年教師實(shí)習(xí)合同協(xié)議樣本:數(shù)學(xué)與科學(xué)教育教師實(shí)習(xí)協(xié)議
- 2025遼寧大連長(zhǎng)興控股集團(tuán)有限公司及所屬公司招聘9人筆試參考題庫(kù)附帶答案詳解
- 教學(xué)技術(shù)與藝術(shù)知到智慧樹(shù)章節(jié)測(cè)試課后答案2024年秋西南大學(xué)
- 健美操知到智慧樹(shù)章節(jié)測(cè)試課后答案2024年秋武漢學(xué)院
- 2025寧夏中匯化工有限公司招聘8人筆試參考題庫(kù)附帶答案詳解
- 2025中國(guó)建材集團(tuán)有限公司招聘14人筆試參考題庫(kù)附帶答案詳解
- 2024遼寧盤(pán)錦市政建設(shè)集團(tuán)社會(huì)招聘31人查看職位筆試參考題庫(kù)附帶答案詳解
- 2025年上半年六盤(pán)水六枝特區(qū)事業(yè)單位及招考易考易錯(cuò)模擬試題(共500題)試卷后附參考答案
- 2025年上半年保山市消防救援支隊(duì)防火監(jiān)督科招聘消防文員4名易考易錯(cuò)模擬試題(共500題)試卷后附參考答案
- 售后服務(wù)流程圖
- 建筑地基處理技術(shù)規(guī)范JGJ79-2012
- 印象主義、后印象主義課件
- 《中華傳統(tǒng)文化》第1課-炎黃始-華夏悠遠(yuǎn)教學(xué)課件
- 日常監(jiān)督檢查表
- 隊(duì)列訓(xùn)練教程ppt課件(PPT 86頁(yè))
- 第三章-農(nóng)村公共管理組織課件
- 注塑員工培訓(xùn)
- JMP操作簡(jiǎn)要培訓(xùn)
- 勝利油田壓驅(qū)技術(shù)工藝研究進(jìn)展及下步工作方向
- 研究生復(fù)試匯報(bào)ppt
評(píng)論
0/150
提交評(píng)論