<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>演讲与分享 - Concordia AI</title>
	<atom:link href="https://concordia-ai.com/zh-hans/category/speaking-2/feed/" rel="self" type="application/rss+xml" />
	<link>https://concordia-ai.com/zh-hans/</link>
	<description>Guiding the governance of AI for a long and flourishing future</description>
	<lastBuildDate>Mon, 01 Dec 2025 02:59:36 +0000</lastBuildDate>
	<language>zh-Hans</language>
	<sy:updatePeriod>
	hourly	</sy:updatePeriod>
	<sy:updateFrequency>
	1	</sy:updateFrequency>
	<generator>https://wordpress.org/?v=6.8.5</generator>

<image>
	<url>https://concordia-ai.com/wp-content/uploads/2025/06/cropped-Favicon-32x32.png</url>
	<title>演讲与分享 - Concordia AI</title>
	<link>https://concordia-ai.com/zh-hans/</link>
	<width>32</width>
	<height>32</height>
</image> 
	<item>
		<title>2025“AI向善”全球峰会：安远AI谢旻希解析前沿AI四大风险与治理路径</title>
		<link>https://concordia-ai.com/zh-hans/2025ai%e5%90%91%e5%96%84%e5%85%a8%e7%90%83%e5%b3%b0%e4%bc%9a%ef%bc%9a%e5%ae%89%e8%bf%9cai%e8%b0%a2%e6%97%bb%e5%b8%8c%e8%a7%a3%e6%9e%90%e5%89%8d%e6%b2%bfai%e5%9b%9b%e5%a4%a7%e9%a3%8e/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=2025ai%25e5%2590%2591%25e5%2596%2584%25e5%2585%25a8%25e7%2590%2583%25e5%25b3%25b0%25e4%25bc%259a%25ef%25bc%259a%25e5%25ae%2589%25e8%25bf%259cai%25e8%25b0%25a2%25e6%2597%25bb%25e5%25b8%258c%25e8%25a7%25a3%25e6%259e%2590%25e5%2589%258d%25e6%25b2%25bfai%25e5%259b%259b%25e5%25a4%25a7%25e9%25a3%258e</link>
		
		<dc:creator><![CDATA[Concordia AI]]></dc:creator>
		<pubDate>Thu, 20 Nov 2025 12:32:04 +0000</pubDate>
				<category><![CDATA[演讲与分享]]></category>
		<guid isPermaLink="false">https://concordia-ai.com/?p=1185</guid>

					<description><![CDATA[<p>导言 2025年7月10日，安远AI创始人兼CEO谢旻希出席了由联合国和国际电信联盟（ITU）共同组织的“AI&#8230;</p>
<p>The post <a href="https://concordia-ai.com/zh-hans/2025ai%e5%90%91%e5%96%84%e5%85%a8%e7%90%83%e5%b3%b0%e4%bc%9a%ef%bc%9a%e5%ae%89%e8%bf%9cai%e8%b0%a2%e6%97%bb%e5%b8%8c%e8%a7%a3%e6%9e%90%e5%89%8d%e6%b2%bfai%e5%9b%9b%e5%a4%a7%e9%a3%8e/">2025“AI向善”全球峰会：安远AI谢旻希解析前沿AI四大风险与治理路径</a> first appeared on <a href="https://concordia-ai.com/zh-hans/">Concordia AI</a>.</p>]]></description>
										<content:encoded><![CDATA[<p><span style="color: #000080;">导言</span></p>
<p>2025年7月10日，安远AI创始人兼CEO谢旻希出席了由联合国和国际电信联盟（ITU）共同组织的“<span style="color: #000080;"><strong>AI向善”全球峰会（AI for Good Summit）“前沿AI治理：从原则到实践”圆桌讨论</strong></span>，探讨前沿人工智能所带来的风险及其治理方案。此次圆桌讨论汇聚了全球人工智能领域的顶尖专家，包括清华大学智能产业研究院院长张亚勤院士、欧盟委员会人工智能顾问Juha Heikkilä、美国前沿模型论坛负责人Chris Meserole，以及来自AWS、Signal等科技企业的企业高管代表。台下观众包括数百位横跨政府、科技、政策与社会创新的各界全球领袖。<span style="color: #000080;"><strong>谢旻希指出AI正面临滥用、意外故障、潜在失控与系统性社会冲击四大风险</strong></span>，其风险治理的先决条件之一是识别和理解前沿AI具有的潜在风险。谢旻希提出“<span style="color: #000080;"><strong>部署前—部署后</strong></span>”两步走路径：既要落实部署前登记与审计，也要建立标识框架推动透明监管与追溯问责，并分享了中国在生成式AI领域的监管实践。他呼吁，<span style="color: #000080;"><strong>防范和管控AI安全风险，需要构建多方协同的治理框架</strong></span>，明确不可接受结果与预警指标，将公民意见与国际共识纳入决策，以避免由少数人冒险导致的系统性危害。</p>
<p>关于管理前沿人工风险的治理模式，谢旻希强调，为了确保人工智能真正造福人类，我们有必要借鉴安全攸关领域的风险管理经验，其先决条件之一是要识别和理解潜在风险。安远AI此前参与的《<a href="https://mp.weixin.qq.com/s?__biz=Mzg4NTgxNjEwMg==&amp;mid=2247501328&amp;idx=1&amp;sn=1ff03d99c0d837eed950c52b64750dca&amp;scene=21#wechat_redirect">国际人工智能安全报告</a>》以及<a href="https://mp.weixin.qq.com/s?__biz=Mzg4NTgxNjEwMg==&amp;mid=2247501804&amp;idx=1&amp;sn=12df5d5bb36512de7fb4dc5d417703c1&amp;scene=21#wechat_redirect">最新相关研究</a>指出，目前前沿人工智能在全球范围内的广泛应用暴露出了四大方面的安全风险挑战：</p>
<p><span style="color: #333399;"><strong>一、滥用</strong></span>：前沿人工智能已经在编程、科学推理、病毒学等两用领域达到专家级水平，为<span style="color: #333399;"><strong>恶意行为者发动复杂网络攻击或设计危险病原体降低了技术门槛</strong></span>。也有研究证实，AI智能体已具备利用零日漏洞的能力。针对这类恶意滥用人工智能的风险，一项制度性的应对办法是向防御方提供早期和优先的访问权限，以便提前研发应对措施，确保系统的防御能力始终先于新兴威胁一步。</p>
<p><span style="color: #333399;"><strong>二、意外事故与故障</strong></span>：聊天机器人在日常对话中表现出的“幻觉”现象无伤大雅，但在医疗诊断等高风险场景下，其幻觉却可能酿成严重伤害。针对这类意外风险，建议<span style="color: #333399;"><strong>明确将特定应用场景归入高风险类别</strong></span>，并施加强有力的<strong><span style="color: #333399;">监管措施</span></strong>保障用户安全。例如，北京市政府已出台监管办法，禁止人工智能系统自动生成<strong><span style="color: #333399;">医用处方</span></strong>。</p>
<p><span style="color: #333399;"><strong>三、失控</strong></span>：正如Geoffrey Hinton在本届峰会所言，目前尚无可靠机制能确保<span style="color: #333399;"><strong>数字超级智能始终在人类掌控之下</strong></span>。已有迹象显示，在某些条件下，前沿模型可能试图欺骗用户或逃脱控制，虽然现阶段人工智能模型还很难绕过客户身份验证（KYC）等安全验证机制，无法在现实环境中自我复制。<span style="color: #333399;"><strong>针对这类高危风险，即便现阶段其实现概率不高或不详，也仍有必要设立预防措施</strong></span>，例如制定危机应对预案，例如开发可靠的系统“开关”等。</p>
<p><span style="color: #333399;"><strong>四、社会层面的系统性风险</strong></span>： 随着通用型人工智能在许多高价值任务上的表现超过人类，劳动力市场受到的来自人工智能的冲击势必逐渐加深。并且，由于这种渐进的影响往往并不集中，任何单一组织都无法独立应对这样<span style="color: #333399;"><strong>广泛的外部性</strong></span>。针对这类系统性风险，我们必须建立多利益相关方的合作机制，来共同管理人工智能催生的</p>
<p>社会层面的冲击与转型，确保无论是地区还是全球范围内不让任何人掉队。这也正是ITU平台的重要价值所在。</p>
<p><img decoding="async" class="alignnone size-full wp-image-1187" src="https://concordia-ai.com/wp-content/uploads/2025/11/Screenshot-2025-11-20-at-12.17.03-PM.png" alt="" width="618" height="292" srcset="https://concordia-ai.com/wp-content/uploads/2025/11/Screenshot-2025-11-20-at-12.17.03-PM.png 618w, https://concordia-ai.com/wp-content/uploads/2025/11/Screenshot-2025-11-20-at-12.17.03-PM-300x142.png 300w, https://concordia-ai.com/wp-content/uploads/2025/11/Screenshot-2025-11-20-at-12.17.03-PM-150x71.png 150w" sizes="(max-width: 618px) 100vw, 618px" /></p>
<p>针对上述各类风险，谢旻希提出，人工智能治理亟需更强有力的措施。当前，针对人工智能的监管比食品安全还少，这在通用型人工智能可能造就深远影响的今天显得尤其不足。未来几年应重点推动两方面的人工智能安全治理举措：一是<strong><span style="color: #333399;">部署前的模型登记与审计</span></strong>，以许可框架确保通用型人工智能在进入市场前经过安全评估与备案，并且可随着认知加深不断优化；二是<strong><span style="color: #333399;">部署后的透明度机制</span></strong>，包括建立生成合成内容标识体系及未来的<span style="color: #333399;"><strong>AI智能体身份标识</strong></span>，确保人工智能系统可追溯、可问责。</p>
<p>展望未来，治理人工智能安全风险更关键的一步，在于建立各方协同的治理框架：<span style="color: #333399;"><strong>明确人工智能系统可能带来的“不可接受结果”或红线，并建立清晰的早期预警指标</strong></span>。人工智能风险阈值的界定不应仅由行业内部决定。全球金融危机和环境灾害已经给过我们历史教训，如果放由少数精英决策者铤而走险，弱势群体所受的冲击只可能更大，甚至可能酿成波及全球的系统性危害。因此，在下一阶段的人工智能治理问题上，凝聚共识、划定边界并强化问责，将是最需要落实的实质性举措。</p><p>The post <a href="https://concordia-ai.com/zh-hans/2025ai%e5%90%91%e5%96%84%e5%85%a8%e7%90%83%e5%b3%b0%e4%bc%9a%ef%bc%9a%e5%ae%89%e8%bf%9cai%e8%b0%a2%e6%97%bb%e5%b8%8c%e8%a7%a3%e6%9e%90%e5%89%8d%e6%b2%bfai%e5%9b%9b%e5%a4%a7%e9%a3%8e/">2025“AI向善”全球峰会：安远AI谢旻希解析前沿AI四大风险与治理路径</a> first appeared on <a href="https://concordia-ai.com/zh-hans/">Concordia AI</a>.</p>]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>安远AI参与2025年SCAI新加坡人工智能会议，会议发布《新加坡共识》建立全球AI安全研究合作新框架</title>
		<link>https://concordia-ai.com/zh-hans/%e5%ae%89%e8%bf%9cai%e5%8f%82%e4%b8%8e2025%e5%b9%b4scai%e6%96%b0%e5%8a%a0%e5%9d%a1%e4%ba%ba%e5%b7%a5%e6%99%ba%e8%83%bd%e4%bc%9a%e8%ae%ae%ef%bc%8c%e4%bc%9a%e8%ae%ae%e5%8f%91%e5%b8%83%e3%80%8a%e6%96%b0/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=%25e5%25ae%2589%25e8%25bf%259cai%25e5%258f%2582%25e4%25b8%258e2025%25e5%25b9%25b4scai%25e6%2596%25b0%25e5%258a%25a0%25e5%259d%25a1%25e4%25ba%25ba%25e5%25b7%25a5%25e6%2599%25ba%25e8%2583%25bd%25e4%25bc%259a%25e8%25ae%25ae%25ef%25bc%258c%25e4%25bc%259a%25e8%25ae%25ae%25e5%258f%2591%25e5%25b8%2583%25e3%2580%258a%25e6%2596%25b0</link>
		
		<dc:creator><![CDATA[Concordia AI]]></dc:creator>
		<pubDate>Sat, 01 Nov 2025 17:10:39 +0000</pubDate>
				<category><![CDATA[演讲与分享]]></category>
		<guid isPermaLink="false">https://concordia-ai.com/?p=1257</guid>

					<description><![CDATA[<p>导言  2025年4月26日，《新加坡共识：全球人工智能安全研究优先事项》在新加坡人工智能会议：人工智能安全国&#8230;</p>
<p>The post <a href="https://concordia-ai.com/zh-hans/%e5%ae%89%e8%bf%9cai%e5%8f%82%e4%b8%8e2025%e5%b9%b4scai%e6%96%b0%e5%8a%a0%e5%9d%a1%e4%ba%ba%e5%b7%a5%e6%99%ba%e8%83%bd%e4%bc%9a%e8%ae%ae%ef%bc%8c%e4%bc%9a%e8%ae%ae%e5%8f%91%e5%b8%83%e3%80%8a%e6%96%b0/">安远AI参与2025年SCAI新加坡人工智能会议，会议发布《新加坡共识》建立全球AI安全研究合作新框架</a> first appeared on <a href="https://concordia-ai.com/zh-hans/">Concordia AI</a>.</p>]]></description>
										<content:encoded><![CDATA[<p><span style="color: #000080;"><b>导言</b></span><span data-ccp-props="{&quot;201341983&quot;:2,&quot;335551550&quot;:6,&quot;335551620&quot;:6,&quot;335559739&quot;:0,&quot;335559740&quot;:384}"> </span></p>
<p><span data-contrast="none">2025</span><span data-contrast="none">年</span><span data-contrast="none">4</span><span data-contrast="none">月</span><span data-contrast="none">26</span><span data-contrast="none">日，</span><span style="color: #000080;">《<b>新加坡共识：全球人工智能安全研究优先事项</b>》在<b>新加坡人工智能会议：人工智能安全国际科学交流会（</b><b>Singapore Conference on AI, SCAI) </b></span><span data-contrast="none">上正式发布。由全球顶尖学术机构的</span><span data-contrast="none">100</span><span data-contrast="none">多位专家共同制定，</span><span style="color: #000080;"><b>明确提出了风险评估、安全开发和可控性三大研究优先领域，旨在通过国际协作共建</b><b>AI</b></span><b><span data-contrast="none"><span style="color: #000080;">安全技术框架</span>。</span></b><span data-ccp-props="{&quot;201341983&quot;:2,&quot;335551550&quot;:6,&quot;335551620&quot;:6,&quot;335559739&quot;:0,&quot;335559740&quot;:384}"> </span><span data-ccp-props="{&quot;201341983&quot;:2,&quot;335551550&quot;:6,&quot;335551620&quot;:6,&quot;335559739&quot;:0,&quot;335559740&quot;:384}"> </span></p>
<p><span data-contrast="none">继</span><a href="https://mp.weixin.qq.com/s?__biz=Mzg4NTgxNjEwMg==&amp;mid=2247494865&amp;idx=1&amp;sn=8ab90903e2c082203b1f33749c43a1ca&amp;scene=21#wechat_redirect"><span data-contrast="none">2023</span><span data-contrast="none">年首次新加坡人工智能会议后</span></a><span data-contrast="none">，安远</span><span data-contrast="none">AI</span><span data-contrast="none">第二次受邀出席并参与了《新加坡共识》初稿的反馈和建议工作。安远</span><span data-contrast="none">AI</span><span data-contrast="none">期待《新加坡共识》能成为加速推进全球</span><span data-contrast="none">AI</span><span data-contrast="none">安全行动的重要契机，共同构建更加安全、可靠的人工智能生态体系。</span><span data-ccp-props="{&quot;201341983&quot;:2,&quot;335551550&quot;:6,&quot;335551620&quot;:6,&quot;335559739&quot;:0,&quot;335559740&quot;:384}"> </span></p>
<p><img loading="lazy" decoding="async" class="alignnone size-full wp-image-1264" src="https://concordia-ai.com/wp-content/uploads/2025/11/1.png" alt="" width="941" height="525" srcset="https://concordia-ai.com/wp-content/uploads/2025/11/1.png 941w, https://concordia-ai.com/wp-content/uploads/2025/11/1-300x167.png 300w, https://concordia-ai.com/wp-content/uploads/2025/11/1-150x84.png 150w, https://concordia-ai.com/wp-content/uploads/2025/11/1-768x428.png 768w" sizes="auto, (max-width: 941px) 100vw, 941px" /></p>
<p><img loading="lazy" decoding="async" class="alignnone size-full wp-image-1262" src="https://concordia-ai.com/wp-content/uploads/2025/11/2-1.png" alt="" width="941" height="447" srcset="https://concordia-ai.com/wp-content/uploads/2025/11/2-1.png 941w, https://concordia-ai.com/wp-content/uploads/2025/11/2-1-300x143.png 300w, https://concordia-ai.com/wp-content/uploads/2025/11/2-1-150x71.png 150w, https://concordia-ai.com/wp-content/uploads/2025/11/2-1-768x365.png 768w" sizes="auto, (max-width: 941px) 100vw, 941px" /></p>
<p><span class="TextRun SCXW104034613 BCX0" data-contrast="none"><span class="NormalTextRun SCXW104034613 BCX0">4</span></span><span class="TextRun SCXW104034613 BCX0" lang="ZH-CN" xml:lang="ZH-CN" data-contrast="none"><span class="NormalTextRun SCXW104034613 BCX0">月</span></span><span class="TextRun SCXW104034613 BCX0" data-contrast="none"><span class="NormalTextRun SCXW104034613 BCX0">26</span></span><span class="TextRun SCXW104034613 BCX0" lang="ZH-CN" xml:lang="ZH-CN" data-contrast="none"><span class="NormalTextRun SCXW104034613 BCX0">日，第二届新加坡人工智能会议</span></span><span class="TextRun SCXW104034613 BCX0" lang="ZH-CN" xml:lang="ZH-CN" data-contrast="none"><span class="NormalTextRun SCXW104034613 BCX0">：人工智能安全国际科学交流会顺利举办。</span></span><span class="TextRun SCXW104034613 BCX0" lang="ZH-CN" xml:lang="ZH-CN" data-contrast="none"><span class="NormalTextRun SCXW104034613 BCX0">会议吸引了麻省理工学院、斯坦福大学、清华大学和中国科学院等学术机构的专家，以及来自</span></span><span class="TextRun SCXW104034613 BCX0" data-contrast="none"><span class="NormalTextRun SCXW104034613 BCX0">OpenAI</span><span class="NormalTextRun SCXW104034613 BCX0">、</span><span class="NormalTextRun SCXW104034613 BCX0">Anthropic</span><span class="NormalTextRun SCXW104034613 BCX0">、</span><span class="NormalTextRun SCXW104034613 BCX0">Google DeepMind</span><span class="NormalTextRun SCXW104034613 BCX0">、</span><span class="NormalTextRun SCXW104034613 BCX0">xAI</span></span><span class="TextRun SCXW104034613 BCX0" lang="ZH-CN" xml:lang="ZH-CN" data-contrast="none"><span class="NormalTextRun SCXW104034613 BCX0">和</span></span><span class="TextRun SCXW104034613 BCX0" data-contrast="none"><span class="NormalTextRun SCXW104034613 BCX0">Meta</span></span><span class="TextRun SCXW104034613 BCX0" lang="ZH-CN" xml:lang="ZH-CN" data-contrast="none"><span class="NormalTextRun SCXW104034613 BCX0">等科技公司的研究人员参与。</span></span><span class="TextRun MacChromeBold SCXW104034613 BCX0" lang="ZH-CN" style="color: #000080;" xml:lang="ZH-CN" data-contrast="none"><span class="NormalTextRun SCXW104034613 BCX0">会议同时促成了《新加坡共识：全球人工智能安全研究优先事项》的发布。</span></span></p>
<p><span class="EOP SCXW104034613 BCX0" data-ccp-props="{&quot;201341983&quot;:2,&quot;335551550&quot;:6,&quot;335551620&quot;:6,&quot;335559739&quot;:0,&quot;335559740&quot;:384}"><img loading="lazy" decoding="async" class="alignnone size-full wp-image-1260" src="https://concordia-ai.com/wp-content/uploads/2025/11/3.png" alt="" width="941" height="622" srcset="https://concordia-ai.com/wp-content/uploads/2025/11/3.png 941w, https://concordia-ai.com/wp-content/uploads/2025/11/3-300x198.png 300w, https://concordia-ai.com/wp-content/uploads/2025/11/3-150x99.png 150w, https://concordia-ai.com/wp-content/uploads/2025/11/3-768x508.png 768w" sizes="auto, (max-width: 941px) 100vw, 941px" /></span></p>
<p><span data-contrast="none">大会</span><b><span data-contrast="none">项目委员会汇聚了全球顶尖学者</span></b><span data-contrast="none">，包括加州大学伯克利分校的</span><span data-contrast="none">Dawn Song</span><span data-contrast="none">教授和</span><span data-contrast="none">Stuart Russell</span><span data-contrast="none">教授、清华大学的薛澜教授和张亚勤教授、南洋理工大学的</span><span data-contrast="none">Luke Ong</span><span data-contrast="none">教授、麻省理工学院及未来生命研究所的</span><span data-contrast="none">Max Tegmark</span><span data-contrast="none">教授、蒙特利尔大学</span><span data-contrast="none">Mila</span><span data-contrast="none">研究院的</span><span data-contrast="none">Tegan Maharaj</span><span data-contrast="none">博士和</span><span data-contrast="none">Yoshua Bengio</span><span data-contrast="none">教授。</span><span data-ccp-props="{&quot;201341983&quot;:2,&quot;335551550&quot;:6,&quot;335551620&quot;:6,&quot;335559739&quot;:0,&quot;335559740&quot;:384}"> </span></p>
<p><span data-contrast="none">安远</span><span data-contrast="none">AI CEO</span><span data-contrast="none">谢旻希与国际</span><span data-contrast="none">AI</span><span data-contrast="none">治理主管吴君仪受邀出席了本次会议，并积极参与了《新加坡共识》初稿的反馈和建议工作。继</span><a href="https://mp.weixin.qq.com/s?__biz=Mzg4NTgxNjEwMg==&amp;mid=2247494865&amp;idx=1&amp;sn=8ab90903e2c082203b1f33749c43a1ca&amp;scene=21#wechat_redirect"><span data-contrast="none">2023</span><span data-contrast="none">年</span><span data-contrast="none">12</span><span data-contrast="none">月首次受邀参与后</span></a><span data-contrast="none">，</span><span data-contrast="none">安远</span><span data-contrast="none">AI</span><span data-contrast="none">再次受邀出席</span><span data-contrast="none">。</span><span data-ccp-props="{&quot;201341983&quot;:2,&quot;335551550&quot;:6,&quot;335551620&quot;:6,&quot;335559739&quot;:0,&quot;335559740&quot;:384}"> </span></p>
<p><span data-contrast="none">《新加坡共识》由全球</span><span data-contrast="none">100</span><span data-contrast="none">多位顶尖专家共同推动，是识别通用型先进人工智能系统潜在风险、推动技术解决方案的重要里程碑，呼吁全球研究者围绕以下</span><b><span data-contrast="none">三大关键领域</span></b><span data-contrast="none">展开深入协作：</span><span data-ccp-props="{&quot;201341983&quot;:2,&quot;335551550&quot;:6,&quot;335551620&quot;:6,&quot;335559739&quot;:0,&quot;335559740&quot;:384}"> </span></p>
<p><img loading="lazy" decoding="async" class="alignnone size-full wp-image-1258" src="https://concordia-ai.com/wp-content/uploads/2025/11/4.png" alt="" width="941" height="802" srcset="https://concordia-ai.com/wp-content/uploads/2025/11/4.png 941w, https://concordia-ai.com/wp-content/uploads/2025/11/4-300x256.png 300w, https://concordia-ai.com/wp-content/uploads/2025/11/4-150x128.png 150w, https://concordia-ai.com/wp-content/uploads/2025/11/4-768x655.png 768w" sizes="auto, (max-width: 941px) 100vw, 941px" /></p>
<ul>
<li aria-setsize="-1" data-leveltext="" data-font="Symbol" data-listid="1" data-list-defn-props="{&quot;335552541&quot;:1,&quot;335559685&quot;:720,&quot;335559991&quot;:360,&quot;469769226&quot;:&quot;Symbol&quot;,&quot;469769242&quot;:[8226],&quot;469777803&quot;:&quot;left&quot;,&quot;469777804&quot;:&quot;&quot;,&quot;469777815&quot;:&quot;multilevel&quot;}" data-aria-posinset="1" data-aria-level="1"><b><span data-contrast="none">风险评估</span></b><span data-contrast="none">：旨在理解潜在危害的严重性和发生可能性，为关键研发和部署决策提供基础。研究方向包括开发衡量</span><span data-contrast="none">AI</span><span data-contrast="none">系统影响的方法，提升计量学能力确保测量结果的准确性和可重复性，以及构建支持第三方审计的机制，实现独立验证。</span><span data-ccp-props="{&quot;201341983&quot;:2,&quot;335551550&quot;:6,&quot;335551620&quot;:6,&quot;335559739&quot;:0,&quot;335559740&quot;:384}"> </span></li>
</ul>
<ul>
<li aria-setsize="-1" data-leveltext="" data-font="Symbol" data-listid="1" data-list-defn-props="{&quot;335552541&quot;:1,&quot;335559685&quot;:720,&quot;335559991&quot;:360,&quot;469769226&quot;:&quot;Symbol&quot;,&quot;469769242&quot;:[8226],&quot;469777803&quot;:&quot;left&quot;,&quot;469777804&quot;:&quot;&quot;,&quot;469777815&quot;:&quot;multilevel&quot;}" data-aria-posinset="2" data-aria-level="1"><b><span data-contrast="none">安全开发</span></b><span data-contrast="none">：致力于从设计层面确保</span><span data-contrast="none">AI</span><span data-contrast="none">系统值得信赖、可靠且安全，增强人们对</span><span data-contrast="none">AI</span><span data-contrast="none">创新的信心。遵循经典安全工程框架，研究方向包括明确</span><span data-contrast="none">AI</span><span data-contrast="none">系统的目标行为规范，设计符合规范的系统，以及验证系统是否满足要求。</span><span data-ccp-props="{&quot;201341983&quot;:2,&quot;335551550&quot;:6,&quot;335551620&quot;:6,&quot;335559739&quot;:0,&quot;335559740&quot;:384}"> </span></li>
</ul>
<ul>
<li aria-setsize="-1" data-leveltext="" data-font="Symbol" data-listid="1" data-list-defn-props="{&quot;335552541&quot;:1,&quot;335559685&quot;:720,&quot;335559991&quot;:360,&quot;469769226&quot;:&quot;Symbol&quot;,&quot;469769242&quot;:[8226],&quot;469777803&quot;:&quot;left&quot;,&quot;469777804&quot;:&quot;&quot;,&quot;469777815&quot;:&quot;multilevel&quot;}" data-aria-posinset="3" data-aria-level="1"><b><span data-contrast="none">可控性</span></b><span data-contrast="none">：通过反馈调节系统行为，确保</span><span data-contrast="none">AI</span><span data-contrast="none">在面对干扰或不确定性时仍能实现预期目标。研究方向包括开发监测与干预机制，拓展监测范围至更广泛的</span><span data-contrast="none">AI</span><span data-contrast="none">生态系统，以及增强社会基础设施适应</span><span data-contrast="none">AI</span><span data-contrast="none">带来的变革，提升整体社会韧性。</span><span data-ccp-props="{&quot;201341983&quot;:2,&quot;335551550&quot;:6,&quot;335551620&quot;:6,&quot;335559739&quot;:0,&quot;335559740&quot;:384}"> </span></li>
</ul>
<p><b><span data-contrast="none">清华大学文科资深教授薛澜在评价这一共识时表示</span></b><span data-contrast="none">：&#8221;</span><span data-contrast="none">在当前地缘政治日益碎片化的背景下，这份关于</span><span data-contrast="none">AI</span><span data-contrast="none">安全前沿研究的全面汇总，体现了全球社会在构建更安全</span><span data-contrast="none">AI</span><span data-contrast="none">未来方面的共同承诺。</span><span data-contrast="none">&#8220;</span><span data-ccp-props="{&quot;201341983&quot;:2,&quot;335551550&quot;:6,&quot;335551620&quot;:6,&quot;335559739&quot;:0,&quot;335559740&quot;:384}"> </span></p>
<p><b><span data-contrast="none">安远</span></b><b><span data-contrast="none">AI</span></b><b><span data-contrast="none">在</span></b><b><span data-contrast="none">2024</span></b><b><span data-contrast="none">年更新发布的《</span></b><a href="https://mp.weixin.qq.com/s?__biz=Mzg4NTgxNjEwMg==&amp;mid=2247500897&amp;idx=1&amp;sn=425c1a8647aa1b9bfa2de07dcfbf09d6&amp;scene=21&amp;token=2119308986&amp;lang=zh_CN&amp;poc_token=HFP4JmijfC3ntqrmFMCReAWwsB9KV1tZjKRyooIc#wechat_redirect"><b><span data-contrast="none">中国人工智能安全全景报告</span></b></a><b><span data-contrast="none">》也显示，中国研究人员正积极贡献全球</span></b><b><span data-contrast="none">AI</span></b><b><span data-contrast="none">安全科学研究和治理体系构建。</span></b><span data-contrast="none">AI</span><span data-contrast="none">安全是各国面临的共同挑战和机遇，安远</span><span data-contrast="none">AI</span><span data-contrast="none">期待《新加坡共识》能成为加速推进全球</span><span data-contrast="none">AI</span><span data-contrast="none">安全行动的重要契机，</span><b><span data-contrast="none">共同构建更加安全、可靠的人工智能生态体系</span></b><b><span data-contrast="none">。</span></b><span data-ccp-props="{&quot;201341983&quot;:2,&quot;335551550&quot;:6,&quot;335551620&quot;:6,&quot;335559739&quot;:0,&quot;335559740&quot;:384}"> </span><span data-ccp-props="{&quot;201341983&quot;:2,&quot;335551550&quot;:6,&quot;335551620&quot;:6,&quot;335559739&quot;:0,&quot;335559740&quot;:384}"> </span></p>
<p><span data-contrast="none">点击</span><a href="https://aisafetypriorities.org/?sessionid="><span data-contrast="none">链接</span></a><span data-contrast="none">，查看完整共识文件。</span><span data-ccp-props="{&quot;201341983&quot;:2,&quot;335551550&quot;:6,&quot;335551620&quot;:6,&quot;335559739&quot;:0,&quot;335559740&quot;:384}"> </span></p>
<p><span class="EOP SCXW104034613 BCX0" data-ccp-props="{&quot;201341983&quot;:2,&quot;335551550&quot;:6,&quot;335551620&quot;:6,&quot;335559739&quot;:0,&quot;335559740&quot;:384}"> </span></p>
<p>&nbsp;</p><p>The post <a href="https://concordia-ai.com/zh-hans/%e5%ae%89%e8%bf%9cai%e5%8f%82%e4%b8%8e2025%e5%b9%b4scai%e6%96%b0%e5%8a%a0%e5%9d%a1%e4%ba%ba%e5%b7%a5%e6%99%ba%e8%83%bd%e4%bc%9a%e8%ae%ae%ef%bc%8c%e4%bc%9a%e8%ae%ae%e5%8f%91%e5%b8%83%e3%80%8a%e6%96%b0/">安远AI参与2025年SCAI新加坡人工智能会议，会议发布《新加坡共识》建立全球AI安全研究合作新框架</a> first appeared on <a href="https://concordia-ai.com/zh-hans/">Concordia AI</a>.</p>]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>AI安全国际论坛 × 复旦 × 安远AI联合举办“AI欺骗风险与治理国际研讨会</title>
		<link>https://concordia-ai.com/zh-hans/ai%e5%ae%89%e5%85%a8%e5%9b%bd%e9%99%85%e8%ae%ba%e5%9d%9b-x-%e5%a4%8d%e6%97%a6-x-%e5%ae%89%e8%bf%9cai%e8%81%94%e5%90%88%e4%b8%be%e5%8a%9eai%e6%ac%ba%e9%aa%97%e9%a3%8e%e9%99%a9/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=ai%25e5%25ae%2589%25e5%2585%25a8%25e5%259b%25bd%25e9%2599%2585%25e8%25ae%25ba%25e5%259d%259b-x-%25e5%25a4%258d%25e6%2597%25a6-x-%25e5%25ae%2589%25e8%25bf%259cai%25e8%2581%2594%25e5%2590%2588%25e4%25b8%25be%25e5%258a%259eai%25e6%25ac%25ba%25e9%25aa%2597%25e9%25a3%258e%25e9%2599%25a9</link>
		
		<dc:creator><![CDATA[Concordia AI]]></dc:creator>
		<pubDate>Tue, 29 Jul 2025 11:44:23 +0000</pubDate>
				<category><![CDATA[演讲与分享]]></category>
		<guid isPermaLink="false">https://concordia-ai.com/?p=1165</guid>

					<description><![CDATA[<p>导读 7月29日，AI安全国际论坛 (SAIF) 、复旦大学计算与智能创新学院以及安远AI共同举办研讨会，聚焦&#8230;</p>
<p>The post <a href="https://concordia-ai.com/zh-hans/ai%e5%ae%89%e5%85%a8%e5%9b%bd%e9%99%85%e8%ae%ba%e5%9d%9b-x-%e5%a4%8d%e6%97%a6-x-%e5%ae%89%e8%bf%9cai%e8%81%94%e5%90%88%e4%b8%be%e5%8a%9eai%e6%ac%ba%e9%aa%97%e9%a3%8e%e9%99%a9/">AI安全国际论坛 × 复旦 × 安远AI联合举办“AI欺骗风险与治理国际研讨会</a> first appeared on <a href="https://concordia-ai.com/zh-hans/">Concordia AI</a>.</p>]]></description>
										<content:encoded><![CDATA[<p><span style="color: #000080;">导读</span></p>
<p>7月29日，AI安全国际论坛 (SAIF) 、复旦大学计算与智能创新学院以及安远AI共同举办研讨会，聚焦人工智能欺骗行为引发的风险。</p>
<p><img loading="lazy" decoding="async" class="alignnone size-full wp-image-1174" src="https://concordia-ai.com/wp-content/uploads/2025/11/Screenshot-2025-11-20-at-11.25.40-AM.png" alt="" width="621" height="465" srcset="https://concordia-ai.com/wp-content/uploads/2025/11/Screenshot-2025-11-20-at-11.25.40-AM.png 621w, https://concordia-ai.com/wp-content/uploads/2025/11/Screenshot-2025-11-20-at-11.25.40-AM-300x225.png 300w, https://concordia-ai.com/wp-content/uploads/2025/11/Screenshot-2025-11-20-at-11.25.40-AM-150x112.png 150w" sizes="auto, (max-width: 621px) 100vw, 621px" /></p>
<p><strong><span style="color: #000080;">这是国内外学者在AI欺骗问题上的首次深度交流</span>。</strong>来自中美德加等国的高校学者、科技企业研究人员和一线研究者齐聚一堂，围绕&#8221;<span style="color: #000080;"><strong>AI欺骗所引发的失控风险威胁模型</strong></span>&#8220;和&#8221;失控风险缓解策略&#8221;两大核心议题展开深入讨论。</p>
<p>会议聚焦AI系统可能具备的误导开发者、掩盖危险能力的&#8221;欺骗性行为&#8221;，提出了针对AI开发生命周期训练、评估、部署三个关键阶段的风险阈值设定方案。与会专家就情境感知能力识别、欺骗行为检测、监控机制规避等技术难题进行了深入研讨，并将基于讨论成果联合撰写AI欺骗风险缓解措施报告。</p>
<p>安远AI CEO谢旻希与技术经理段雅文参与研讨</p>
<h3>会议概要</h3>
<p>本次研讨会围绕 “<strong>‘<span style="color: #000080;">AI欺骗’所引发的失控风险威胁模型</span></strong>” 以及 “<span style="color: #000080;"><strong>失控风险缓解策略</strong></span>” 两大核心议题展开。</p>
<p>会议指出，AI系统可能具备误导开发者、掩盖其危险性能力、倾向及行为的能力，且此类行为难以被察觉。</p>
<p><strong>这种“<span style="color: #000080;">欺骗性行为”可能贯穿AI开发生命周期的全过程</span></strong>，具体表现为：在训练阶段操纵训练过程、在评估阶段干扰评估结果、在部署阶段规避监控机制。</p>
<p>为应对上述风险，研讨会提炼出针对训练、评估、部署三个关键阶段设定的三道风险阈值及应对措施。</p>
<p>与会人员重点围绕这三道风险阈值和缓解措施展开了深入讨论。</p>
<h3>阈值一：AI系统具备情境感知能力</h3>
<p>首先，美国加州非盈利组织Truthful AI的研究员James Chua组织与会者研讨了第一道阈值的划分标准及应对措施。</p>
<p>该阈值设定于 <span style="color: #000080;"><strong>AI 系统具备情境感知能力 (Situational Awareness)</strong></span> 的时刻。</p>
<p>该能力具体指<span style="color: #000080;"><strong>AI能够准确判断自身当前是处于训练、评估还是部署阶段，并能利用这一信息思考如何达成人类设定的相应目标</strong></span>。</p>
<p>若发现AI系统越过了此阈值，开发者需<strong>对其在训练和评估阶段可能采取的破坏训练过程、操作评估的行为进行全面测试</strong>。</p>
<p><img loading="lazy" decoding="async" class="alignnone size-full wp-image-1172" src="https://concordia-ai.com/wp-content/uploads/2025/11/Screenshot-2025-11-20-at-11.29.10-AM.png" alt="" width="621" height="466" srcset="https://concordia-ai.com/wp-content/uploads/2025/11/Screenshot-2025-11-20-at-11.29.10-AM.png 621w, https://concordia-ai.com/wp-content/uploads/2025/11/Screenshot-2025-11-20-at-11.29.10-AM-300x225.png 300w, https://concordia-ai.com/wp-content/uploads/2025/11/Screenshot-2025-11-20-at-11.29.10-AM-150x113.png 150w" sizes="auto, (max-width: 621px) 100vw, 621px" /></p>
<h3>阈值二：AI系统能可靠地破坏训练，博弈评估</h3>
<p>复旦大学计算与智能创新学院副研究员，上海创智学院导师潘旭东组织与会人员就第二道阈值的设定及其应对方案进行了研讨。该阈值被设定于AI系统在前期测试中被确认<strong>能可靠地展现出破坏训练过程及操纵评估的能力</strong>。</p>
<p>提出该阈值的研究者认为，若发现系统表现出此类能力，开发者需采取更加强有力的应对措施：比如<span style="color: #000080;"><strong>故意误导AI系统以干扰其对当前开发阶段的判断，在评估过程中引入更多噪声，或引导系统提升其行为的诚实性与透明度</strong></span>。</p>
<p><img loading="lazy" decoding="async" class="alignnone size-full wp-image-1170" src="https://concordia-ai.com/wp-content/uploads/2025/11/Screenshot-2025-11-20-at-11.29.28-AM.png" alt="" width="621" height="413" srcset="https://concordia-ai.com/wp-content/uploads/2025/11/Screenshot-2025-11-20-at-11.29.28-AM.png 621w, https://concordia-ai.com/wp-content/uploads/2025/11/Screenshot-2025-11-20-at-11.29.28-AM-300x200.png 300w, https://concordia-ai.com/wp-content/uploads/2025/11/Screenshot-2025-11-20-at-11.29.28-AM-150x100.png 150w" sizes="auto, (max-width: 621px) 100vw, 621px" /></p>
<h3>阈值三：AI系统能可靠地规避现有的监控机制</h3>
<p>最后，在加拿大魁北克人工智能研究所博士后Sören Mindermann (首份国际AI安全报告技术负责人) 的组织下，与会人员讨论了第三道 (也是最后一道) 阈值：<span style="color: #000080;"><strong>AI系统在大规模部署中能够可靠规避所有现有的监控机制</strong></span>。</p>
<p>该阈值建议：<span style="color: #000080;"><strong>一旦触发此阈值，开发者必须立即停止部署，无论系统是在内部上线还是通过公开API发布</strong></span>。唯有在引入更强大的监控与安全措施，并确保系统无法规避这些机制后，方可恢复部署。</p>
<p><strong>该原则同样适用于开源模型</strong>：一旦模型参数公开，监控责任即转移至使用者；在缺乏同等效力防护措施的情况下，不应运行该系统。</p>
<p><img loading="lazy" decoding="async" class="alignnone size-full wp-image-1168" src="https://concordia-ai.com/wp-content/uploads/2025/11/Screenshot-2025-11-20-at-11.30.30-AM.png" alt="" width="619" height="465" srcset="https://concordia-ai.com/wp-content/uploads/2025/11/Screenshot-2025-11-20-at-11.30.30-AM.png 619w, https://concordia-ai.com/wp-content/uploads/2025/11/Screenshot-2025-11-20-at-11.30.30-AM-300x225.png 300w, https://concordia-ai.com/wp-content/uploads/2025/11/Screenshot-2025-11-20-at-11.30.30-AM-150x113.png 150w" sizes="auto, (max-width: 619px) 100vw, 619px" /></p>
<h3>交流与分享</h3>
<p>参会者针对每个环节的阈值及应对策略展开了深入讨论，比如是否可以跳过对于系统情境感知能力的测试（即阈值一），直接测试系统是否具备欺骗能力（即阈值二）；如何定义第二和第三部分提及的系统能力的“可靠性”；以及是否应强制要求模型内置监控工具，或者延迟模型发布直至监管体系完善。</p>
<p><img loading="lazy" decoding="async" class="alignnone size-full wp-image-1166" src="https://concordia-ai.com/wp-content/uploads/2025/11/Screenshot-2025-11-20-at-11.30.41-AM.png" alt="" width="621" height="464" srcset="https://concordia-ai.com/wp-content/uploads/2025/11/Screenshot-2025-11-20-at-11.30.41-AM.png 621w, https://concordia-ai.com/wp-content/uploads/2025/11/Screenshot-2025-11-20-at-11.30.41-AM-300x224.png 300w, https://concordia-ai.com/wp-content/uploads/2025/11/Screenshot-2025-11-20-at-11.30.41-AM-150x112.png 150w" sizes="auto, (max-width: 621px) 100vw, 621px" /></p>
<p>会议主办方将根据讨论内容，对起草的阈值和缓解措施进行修改，联合与会者一起撰写一份关于人工智能欺骗风险和缓解措施的报告。敬请期待！</p>
<h3>主办方</h3>
<p>复旦大学计算与智能创新学院以原计算机科学技术学院为班底，整合国家示范性软件学院、国家保密学院、密码学院、网络安全学院，纳入大数据研究院、人工智能创新与产业 (AI³)研究院、金融科技研究院、先进计算系统研究院等多个实体科研机构，致力打造全球一流的计算与智能领域研究型学院。</p>
<p>安远AI致力于引领人际关系走向安全、可信、可靠的未来。在AI安全和治理领域。安远AI既是第三方研究和咨询机构，也是目前该领域全国唯一的社会企业。</p>
<p>AI安全国际论坛是一家非盈利组织，致力于推动全球人工智能治理合作，促进中西AI专家的第二轨道对话，共同应对AI带来的极端风险，造福全人类。我们主办AI安全国际对话，并提供研究和咨询服务。</p><p>The post <a href="https://concordia-ai.com/zh-hans/ai%e5%ae%89%e5%85%a8%e5%9b%bd%e9%99%85%e8%ae%ba%e5%9d%9b-x-%e5%a4%8d%e6%97%a6-x-%e5%ae%89%e8%bf%9cai%e8%81%94%e5%90%88%e4%b8%be%e5%8a%9eai%e6%ac%ba%e9%aa%97%e9%a3%8e%e9%99%a9/">AI安全国际论坛 × 复旦 × 安远AI联合举办“AI欺骗风险与治理国际研讨会</a> first appeared on <a href="https://concordia-ai.com/zh-hans/">Concordia AI</a>.</p>]]></content:encoded>
					
		
		
			</item>
	</channel>
</rss>
