How to legislate for AI in an age of uncertainty - FT中文网
登录×
电子邮件/用户名
密码
记住我
请输入邮箱和密码进行绑定操作:
请输入手机号码,通过短信验证(目前仅支持中国大陆地区的手机号):
请您阅读我们的用户注册协议隐私权保护政策,点击下方按钮即视为您接受。
人工智能

How to legislate for AI in an age of uncertainty

We need laws that only kick in once we know the impact of the technology

The writer is professor of law at Penn State

We don’t know the future, nor how artificial intelligence will affect it. Some believe AI will lead to explosive economic growth, others are convinced it will immiserate all but a select few, while some aver that its economic impact will be marginal.

So how do legislators do their job under such uncertainty? Currently they don’t. They have abrogated their responsibilities to promote the health, safety and welfare of voters through inaction, adopting a wait-and-see approach. If they delay too long, the new technology could have already harmed society and generated new billionaires ready to capture future regulatory processes. Yet regulating too early also has risks, inadvertently hampering innovation. 

There is a better way to proceed that allows us to respond proactively to uncertainty. We need adaptive AI laws that detail how to react to each possible future harm or benefit but that don’t kick in until we see how AI is transforming society. Such adaptive AI laws would be passed now and take effect automatically when benchmarks are met to hedge social risks and distribute benefits.  

Adaptive AI laws could take many forms and borrow tools from elsewhere, such as the decision trees used in machine learning. For example, politicians can currently pass one set of laws that take effect if job losses mount, triggering policies like supplemental unemployment benefits and increased taxes on the rich. Another set of acts could be triggered with job growth, such as improved sick leave and fewer corporate subsidies. Benefits and taxes could rise using sliding scales tied to job losses or income inequality. Some responses, such as instituting a universal wage, could activate under numerous scenarios if economic inequality got too bad or if economic growth exploded. 

Such adaptive AI legislation could be applied to a range of other fields. It was initially unknown how harmful social media would be for children. If it emerges that AI has similarly negative effects, adaptive laws could restrict children’s access to it. If AI instead advances mental health goals, more public resources could be allocated. 

It’s easy to assume that AI will enhance our learning and expertise, yet a recent study showed that oncologists were 20 per cent worse at detecting precancerous growths on their own after having relied on AI as a detection aid. Adaptive AI regulation could grapple with the uncertain effects of AI on education across disciplines and ages. 

Many benchmarks for triggering the activation of an adaptive law could draw from reliable governmental data, such as on income distributions, educational attainment and lifespans. Triggering measurements for other topics, such as mental health, would be more complex to gauge and would benefit from bipartisan legislative guidance and monitoring from designated agencies. If partisan disagreement surfaces about whether a triggering event has occurred, courts can fulfil their established role as legal interpreters. 

Adaptive AI laws would provide three main benefits. First, they would empower lawmakers to act now to avoid future problems, to be proactive instead of reactive. This would occur without sacrificing flexibility, because legislators could always change the adaptive regulation in response to technological developments or social change. Second, the structure of adaptive regulation would encourage lawmakers to think more deeply about the different possible paths that AI could take, which one hopes would lead to better policies. Third, adaptive AI regulations would provide a more stable regulatory framework for AI labs, creating legal clarity by informing labs how laws will automatically change depending on their actions. 

This wouldn’t be the first time that legislatures adopted laws that only activate under certain scenarios. For example, states have passed trigger laws contingent on developments related to abortionMedicaid and rent control.

AI labs have already committed to voluntary if-then commitments, pledging to enhance safety measures once AI models have a particular capability. Yet these commitments are nonbinding, not universal and only touch on safety considerations, not how AI will affect society more broadly. 

We can reasonably imagine the different effects AI might have on society, but we can’t predict which path it will take. Adaptive AI laws would allow us to think through how to regulate the technology now without delaying until it’s too late or hurting innovation through having the regulations kick in immediately. To manage the potential AI revolution, law needs one of its own. 

版权声明:本文版权归FT中文网所有,未经允许任何单位或个人不得转载,复制或以任何其他方式使用本文全部或部分,侵权必究。

数据公司:古巴石油储备仅能再撑“15到20天”

由于美国封锁了委内瑞拉的石油交付,并向另一供应国墨西哥施压,运往哈瓦那的原油已枯竭。

制裁见效,俄罗斯石油收入大幅下滑

俄罗斯的能源收入在2025年比上一年下降了五分之一。

Lex专栏:Meta的快速增长说明什么

按市值计算,美国六大科技巨头占标普500指数的比重已超40%。

飞机租赁业高管:到2050年实现零碳飞行是“空中画饼”

AerCap公司的高管凯利表示,没有人愿意为可持续航空燃料支付更高成本。

印度最大IT服务公司负责人:AI不会导致大规模裁员

人工智能技术的采用正帮助外包企业抵消欧美销售放缓的影响。

“节约型家庭”成为日本关键选民

通胀回潮、利率上升冲击的经济环境让食品成本成为头号政治关切,日本政治正在适应该国中间阶层分化的现实。
设置字号×
最小
较小
默认
较大
最大
分享×