<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>Announcement - Concordia AI</title>
	<atom:link href="https://concordia-ai.com/category/announcement/feed/" rel="self" type="application/rss+xml" />
	<link>https://concordia-ai.com</link>
	<description>Guiding the governance of AI for a long and flourishing future</description>
	<lastBuildDate>Tue, 17 Mar 2026 02:36:03 +0000</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>
	hourly	</sy:updatePeriod>
	<sy:updateFrequency>
	1	</sy:updateFrequency>
	<generator>https://wordpress.org/?v=6.8.5</generator>

<image>
	<url>https://concordia-ai.com/wp-content/uploads/2025/06/cropped-Favicon-32x32.png</url>
	<title>Announcement - Concordia AI</title>
	<link>https://concordia-ai.com</link>
	<width>32</width>
	<height>32</height>
</image> 
	<item>
		<title>2025 Q4 Update from our Frontier AI Risk Monitoring Platform</title>
		<link>https://concordia-ai.com/2025-q4-update-from-our-frontier-ai-risk-monitoring-platform/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=2025-q4-update-from-our-frontier-ai-risk-monitoring-platform</link>
		
		<dc:creator><![CDATA[Concordia AI]]></dc:creator>
		<pubDate>Tue, 17 Mar 2026 02:33:17 +0000</pubDate>
				<category><![CDATA[Announcement]]></category>
		<guid isPermaLink="false">https://concordia-ai.com/?p=1397</guid>

					<description><![CDATA[<p>We have released the 2025 Q4 update of our Frontier AI Risk Monitoring Report (2025Q4)! This is the second report since we launched the Frontier AI Risk Monitoring Platform last year. It tracks models from 16 leading developers worldwide for risks in four domains: cyber offense, biological risks, chemical risks, and loss-of-control. This report tracks frontier models released in the&#8230;</p>
<p>The post <a href="https://concordia-ai.com/2025-q4-update-from-our-frontier-ai-risk-monitoring-platform/">2025 Q4 Update from our Frontier AI Risk Monitoring Platform</a> first appeared on <a href="https://concordia-ai.com">Concordia AI</a>.</p>]]></description>
										<content:encoded><![CDATA[<p>We have released the 2025 Q4 update of our <a href="https://airiskmonitor.net/doc/en/report/2025-Q4" rel="">Frontier AI Risk Monitoring Report (2025Q4)</a>! This is the second report since we <a href="https://aisafetychina.substack.com/p/10-key-insights-from-concordia-ais" rel="">launched</a> the <a href="https://airiskmonitor.net/" rel="">Frontier AI Risk Monitoring Platform</a> last year. It tracks models from 16 leading developers worldwide for risks in four domains: cyber offense, biological risks, chemical risks, and loss-of-control.</p>
<p><img decoding="async" class="alignnone size-full wp-image-1398" src="https://concordia-ai.com/wp-content/uploads/2026/03/platform-report-en.png" alt="" width="1270" height="719" srcset="https://concordia-ai.com/wp-content/uploads/2026/03/platform-report-en.png 1270w, https://concordia-ai.com/wp-content/uploads/2026/03/platform-report-en-300x170.png 300w, https://concordia-ai.com/wp-content/uploads/2026/03/platform-report-en-1024x580.png 1024w, https://concordia-ai.com/wp-content/uploads/2026/03/platform-report-en-150x85.png 150w, https://concordia-ai.com/wp-content/uploads/2026/03/platform-report-en-768x435.png 768w" sizes="(max-width: 1270px) 100vw, 1270px" /></p>
<p>This report tracks frontier models released in the fourth quarter of 2025 and synthesizes trends of the full year, offering a comprehensive view of the evolving AI risk landscape.</p>
<p>While Q3 2025 saw sharp rises in Risk Indices, Q4 presents a more nuanced picture: overall risk levels have stabilized, and frontier models show significant safety gains.</p>
<p><em>Note: The Risk Index is a score that reflects the overall risk of a model by combining its Capability Score and Safety Score. Higher Capability Scores and lower Safety Scores results in a higher Risk Index. Details about the methodology and limitations are <a href="https://airiskmonitor.net/doc/en/about#evaluation-methodology" rel="">here</a>.</em></p>
<p>Here are 10 key insights from our latest monitoring data:</p>
<h4 class="header-anchor-post">1. Overall risk indices have stabilized</h4>
<p>In contrast to the previous period, when Risk Indices hit record highs across all domains, Risk Indices for models released in Q4 2025 did not set new records. This suggests a momentary stabilization in the aggregate risk level of frontier models.</p>
<h4 class="header-anchor-post">2. Risk trends diverge significantly across model families</h4>
<p>While the overall trend is stable, individual model families followed distinct trajectories in Q4:</p>
<ul>
<li>Stable Low Risk: The GPT and Claude families maintained consistently low Risk Indices.</li>
<li>Stable High Risk: The DeepSeek family remained stable but at relatively high risk levels.</li>
<li>Risk Reduction: The Doubao, Hunyuan, and MiniMax families saw significant decreases in Risk Indices.</li>
<li>Risk Increase: The Gemini and Kimi families saw increases in specific domains (e.g., Gemini in biological and loss-of-control risks).</li>
</ul>
<h4 class="header-anchor-post">3. Significant Improvement in Safety Scores for Frontier Models</h4>
<p>Safety Scores for models released in Q4 2025 rose significantly compared to the previous quarter, signaling a marked improvement in the safety of new releases. The Doubao, Hunyuan, and MiniMax families demonstrated the most notable gains.</p>
<h4 class="header-anchor-post">4. Open-weight models lag behind proprietary models in cyber and bio capabilities</h4>
<p>Consistent with the previous quarter, open-weight models rival proprietary ones in chemical and loss-of-control capabilities but lag notably in cyber offense and biological capabilities. The gap in the biological domain is widening, approaching a one-year lag.</p>
<h4 class="header-anchor-post">5. Cyberattack capabilities have reached new heights</h4>
<p>Despite stabilizing risk indices, raw capabilities continue to grow. GPT-5.2 (high) achieved a breakthrough score of 94.7 on the <em>CyberSecEval2-VulnerabilityExploit</em> benchmark, indicating exceptional proficiency in identifying and exploiting software vulnerabilities. Claude Opus 4.5 Reasoning topped the <em>WMDP-Cyber</em> benchmark with a score of 90.3.</p>
<h4 class="header-anchor-post">6. Biological capabilities now surpass human experts in key tasks</h4>
<p>Q4 models have crossed critical thresholds in biology. Gemini 3 Pro Preview has surpassed human expert levels in sequence understanding, cloning experiments, and wet lab troubleshooting. This marks a significant milestone in AI’s utility—and potential risk—in the biological domain.</p>
<h4 class="header-anchor-post">7. &#8230;But biological safeguards lag behind capabilities</h4>
<p>The gap between capability and safety is most acute in the biological domain. Despite its superhuman capabilities, Gemini 3 Pro Preview exhibited a refusal rate of only 57.2% for harmful biological queries on the <em>SciKnowEval</em> benchmark, highlighting a concerning safety lag.</p>
<h4 class="header-anchor-post">8. Chemical safety refusal rates have increased widely</h4>
<p>While capability growth in the chemical domain has plateaued, safety has improved. 70% of models released in Q4 exceeded an 80% refusal rate for harmful chemical queries (measured by <em>SOSBench-Chem</em>), representing a strong improvement over previous quarters.</p>
<h4 class="header-anchor-post">9. Jailbreak safeguards have strengthened</h4>
<p>Defense against adversarial attacks has improved. Models released in Q4 showed significantly stronger resistance to jailbreaking on the <em>StrongReject </em>benchmark. The Claude and GPT families lead with high robustness, while the MiniMax family showed the most notable quarter-over-quarter improvement.</p>
<h4 class="header-anchor-post">10. Loss-of-control risks: High awareness, polarized honesty</h4>
<ul>
<li>Situational Awareness (e.g. awareness of whether they are in training or deployment stage): High situational awareness is a necessary condition for loss-of-control; the higher the score, the greater the risk. Most Q4 models scored near or above 80 out of 100 points. In comparison, in the previous quarter, only 2 models scored above 80 points, with the majority falling below 80.</li>
<li>Honesty: Performance is highly uneven. While Claude Opus 4.5 Reasoning achieved a high honesty score of 96.4, other models like Gemini 3 Pro Preview scored as low as 44.7.</li>
</ul>
<p><em>Note: Our current methodology on the loss-of-control is not yet perfect. We plan to improve in the next version.</em></p>
<h2 class="header-anchor-post"><strong>Explore the Data</strong></h2>
<p>These insights only scratch the surface. We invite you to explore the full interactive data, methodology, and model breakdowns on the <a href="https://airiskmonitor.net/" rel="">Frontier AI Risk Monitoring Platform</a>.</p>
<p>For a detailed analysis of these trends, read the <a href="https://airiskmonitor.net/doc/en/report/2025-Q4" rel="">full report</a>.</p>
<div class="subscription-widget-wrap">
<div class="subscription-widget show-subscribe">
<div class="preamble"></div>
</div>
</div><p>The post <a href="https://concordia-ai.com/2025-q4-update-from-our-frontier-ai-risk-monitoring-platform/">2025 Q4 Update from our Frontier AI Risk Monitoring Platform</a> first appeared on <a href="https://concordia-ai.com">Concordia AI</a>.</p>]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Concordia AI 2025 Impact Highlights</title>
		<link>https://concordia-ai.com/concordia-ai-2025-impact-highlights/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=concordia-ai-2025-impact-highlights</link>
		
		<dc:creator><![CDATA[Concordia AI]]></dc:creator>
		<pubDate>Fri, 06 Mar 2026 06:35:41 +0000</pubDate>
				<category><![CDATA[Announcement]]></category>
		<guid isPermaLink="false">https://concordia-ai.com/?p=1387</guid>

					<description><![CDATA[<p>Throughout 2025, frontier AI capabilities advanced rapidly. But the same capabilities that make these systems so useful also introduce new societal risks. Real-world evidence for several of these risks continues to grow—across malicious use, malfunctions, and systemic threats. Against this backdrop, Concordia AI’s mission remains as critical as ever: ensuring that AI is developed and&#8230;</p>
<p>The post <a href="https://concordia-ai.com/concordia-ai-2025-impact-highlights/">Concordia AI 2025 Impact Highlights</a> first appeared on <a href="https://concordia-ai.com">Concordia AI</a>.</p>]]></description>
										<content:encoded><![CDATA[<p>Throughout 2025, frontier AI capabilities advanced rapidly. But the same capabilities that make these systems so useful also introduce new societal risks. Real-world evidence for several of these risks continues to grow—across malicious use, malfunctions, and systemic threats.</p>
<p>Against this backdrop, Concordia AI’s mission remains as critical as ever: ensuring that AI is developed and deployed safely and in alignment with global interests. We advance this mission through research, advisory work with leading AI companies and policymakers, and promotion of international dialogue.</p>
<p>Below are some of our key accomplishments in 2025. We’ve organized this list according to key axes of our work: international convenings, international research and public engagement, and contributing to China’s domestic AI safety and governance landscape. We end with organizational updates. For previous highlights, see our <a href="https://aisafetychina.substack.com/p/concordia-ai-2023-annual-review" rel="">2023</a>, <a href="https://aisafetychina.substack.com/p/concordia-ai-2024-impact-highlights" rel="">2024</a> and <a href="https://aisafetychina.substack.com/p/concordia-ai-2025-mid-year-impact" rel="">mid-2025</a> reports.</p>
<h2 class="header-anchor-post">International convenings</h2>
<ul>
<li><strong>Convening international AI safety dialogues in China, Singapore, and globally</strong>
<ul>
<li>Hosted the <a href="https://aisafetychina.substack.com/p/concordia-ai-holds-the-ai-safety" rel="">AI Safety and Governance Forum at the World AI Conference (WAIC)</a>. This was Concordia AI’s flagship convening of 2025—bringing together around 30 distinguished experts from around the world, including Turing Award winner Yoshua Bengio; United Nations Under-Secretary-General Amandeep Singh Gill; Shanghai AI Lab Director ZHOU Bowen (周伯文); Special Envoy of the President of France for AI Anne Bouverot; Distinguished Professor of computer science at UC Berkeley Stuart Russell; Peng Cheng Laboratory Director GAO Wen (高文). We had 200+ in-person attendees and 14,000+ livestream views. The Forum was covered my multiple media outlets, including <a href="https://www.bloomberg.com/news/articles/2025-07-30/china-prepares-to-unseat-us-in-fight-for-4-8-trillion-ai-market" rel="">Bloomberg</a>, <a href="https://www.wired.com/story/china-artificial-intelligence-policy-laws-race/" rel="">Wired</a>, <a href="https://science.caixin.com/m/2025-07-30/102346902.html" rel="">Caixin</a>, <a href="https://mp.weixin.qq.com/s/EwDrlAveGkMm7NsnqCZi6Q" rel="">IT Times</a>, and <a href="https://techreviewafrica.com/news/2580/un-digital-envoy-concludes-china-visit-advocates-for-inclusive-ai-governance" rel="">Tech Review Africa</a>. We also co-hosted/hosted multiple side events and expert workshops and served as official AI Governance Advisor for WAIC 2025.</li>
<li>Co-hosted two international workshops with the Carnegie Endowment for International Peace, the Oxford Martin School AI Governance Initiative, the Oxford China Policy Lab, Tsinghua University Center for International Security and Strategy (CISS), and Tsinghua University Institute for AI International Governance (I-AIIG). The first workshop focused on “AI Safety as a Collective Challenge”, and was held at the French AI Action Summit in January. The second focused on “Early Warning and Crisis Coordination for Advanced AI” and was held at the World AI Conference in Shanghai in July.<span data-state="closed"><a id="footnote-anchor-1-189963930" class="footnote-anchor" href="https://aisafetychina.substack.com/p/concordia-ai-2025-impact-highlights?utm_source=post-email-title&amp;publication_id=1862105&amp;post_id=189963930&amp;utm_campaign=email-post-title&amp;isFreemail=true&amp;r=6j32gs&amp;triedRedirect=true&amp;utm_medium=email#footnote-1-189963930" target="_self" rel="" data-component-name="FootnoteAnchorToDOM">1</a></span></li>
<li>Co-hosted the AI Safety Forum at the <a href="https://2025.baai.ac.cn/schedule" rel="">Beijing Academy of AI Conference 2025</a> where technical experts from institutions including MIT, Fudan University, Singapore Management University, and Tsinghua University worked to build consensus on AI red lines.</li>
<li>Organised an <a href="https://www.linkedin.com/feed/update/urn:li:activity:7340690669990031360/" rel="">AI Risk Management Workshop</a> on the sidelines of Asia Tech x Singapore (May 2025) with 20+ experts in AI safety — spanning policy, industry, AI assurance and academia — with participants based across Singapore, China, the US, UK, and the EU, with the support of the Infocomm Media Development Authority of Singapore (IMDA).</li>
<li>Co-hosted a “<a href="https://www.linkedin.com/posts/concordia-ai_frontier-ai-is-reshaping-cyber-riskand-activity-7393896490525691904-tLkk" rel="">Frontier AI in Cybersecurity” workshop</a> with Nanyang Technological University CyberSG R&amp;D Programme Office and UC Berkeley RDI which brought together 25 leaders across government, law enforcement agencies, leading AI labs; and organised the <a href="https://luma.com/wyereks8" rel="">AI Governance in Singapore panel</a> at Lorong AI, on the sidelines of the Singapore International Cybersecurity Week 2025.</li>
<li>Co-hosted events at the International Conference on Learning Representations (ICLR) 2025 Singapore: a “Frontier Governance Exchange” with Singapore AI Safety Hub, Lorong AI and Safe AI Forum; an AI safety social attended by 130+ participants and a “Misalignment and Control” workshop with FAR.AI and Singapore AI Safety Hub.</li>
</ul>
</li>
</ul>
<p><img loading="lazy" decoding="async" class="alignnone size-full wp-image-1162" src="https://concordia-ai.com/wp-content/uploads/2025/11/Screenshot-2025-11-20-at-10.41.09-AM.png" alt="" width="843" height="532" srcset="https://concordia-ai.com/wp-content/uploads/2025/11/Screenshot-2025-11-20-at-10.41.09-AM.png 843w, https://concordia-ai.com/wp-content/uploads/2025/11/Screenshot-2025-11-20-at-10.41.09-AM-300x189.png 300w, https://concordia-ai.com/wp-content/uploads/2025/11/Screenshot-2025-11-20-at-10.41.09-AM-150x95.png 150w, https://concordia-ai.com/wp-content/uploads/2025/11/Screenshot-2025-11-20-at-10.41.09-AM-768x485.png 768w" sizes="auto, (max-width: 843px) 100vw, 843px" /></p>
<div class="captioned-image-container" style="text-align: center;">
<figure><figcaption class="image-caption"><em><span style="color: #999999;">Group photo after the WAIC AI Safety and Governance Forum morning session.</span></em></figcaption></figure>
</div>
<ul>
<li><strong>Contributing to and participating in global and multilateral AI governance efforts</strong>
<ul>
<li>Concordia AI CEO Brian TSE (谢旻希) participated in the <a href="https://www.aistandardssummit.org/event/354f4a77-ee25-47e3-8e84-291a55519c0c/programme" rel="">International AI Standards Summit</a> (Seoul, Dec 2–3) and spoke on a panel, as part of the expert delegation recommended by the National Standardization Administration of China.</li>
<li>Brian Tse was invited as a Chinese civil society representative to the French AI Action Summit in the Grand Palais. Invited to a closed-door seminar hosted by the China AI Safety &amp; Development Association (CnAISDA).</li>
<li>Provided <a href="https://www.un.org/global-digital-compact/en/ai-panel-inputs" rel="">written inputs</a> to UN consultations regarding the Independent International Scientific Panel on AI and Global Dialogue on AI.</li>
<li><a href="https://www.youtube.com/watch?v=4XMip8phUn4" rel="">Spoke</a> on the panel “From Principles to Practice—Governing Advanced AI in Action” at the <a href="https://aiforgood.itu.int/summit25/programme/" rel="">AI for Good Summit 2025</a>.</li>
<li>Participated in the International Dialogues on AI Safety (Shanghai) and signed the <a href="https://idais.ai/dialogue/idais-shanghai/" rel="">Shanghai Consensus</a> on “Ensuring Alignment and Human Control of Advanced AI Systems to Safeguard Human Flourishing”.</li>
</ul>
</li>
<li><strong>Global AIxBiosecurity governance: </strong>We contributed to a number of critical global discussions at the intersection of AI and biosecurity:
<ul>
<li>Brian Tse signed the <a href="https://www.nti.org/analysis/articles/statement-on-biosecurity-risks-at-the-convergence-of-ai-and-the-life-sciences/" rel="">Statement on Biosecurity Risks at the Convergence of AI and the Life Sciences</a> along with figures such as Andrew Yao, Yoshua Bengio, and George Church, and presented the Statement during <a href="https://substack.com/redirect/3d6aff06-9bd9-4352-a70f-845f57448efe?j=eyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM" rel="">The Sixth Session of the Working Group on the Strengthening of the Biological Weapons Convention</a>, as part of the <a href="https://www.nti.org/about/programs-projects/project/aixbio-global-forum/" rel="">Global AIxBio Global Forum</a>.</li>
<li>Participated in an <a href="https://www.nti.org/news/nti-at-the-munich-security-conference-reducing-nuclear-and-biological-risks-together/" rel="">AIxBio tabletop exercise at the Munich Security Conference</a> hosted by the Nuclear Threat Initiative, which led to the publication of the report “<a href="https://www.nti.org/events/report-launch-safeguarding-aixbio-capabilities-to-prevent-global-catastrophe/" rel="">Safeguarding Against Global Catastrophe: Risks, Opportunities, and Governance Options at the Intersection of Artificial Intelligence and Biology.</a>”</li>
<li>Presented at a WHO dialogue on AIxBio implications for the <a href="https://www.who.int/groups/technical-advisory-group-on-the-responsible-use-of-the-life-sciences-and-dual-use-research-(tag-ruls-dur)" rel="">Technical Advisory Group on the Responsible Use of the Life Sciences and Dual-Use Research</a>.</li>
<li>Co-developed and endorsed the “<a href="https://unodaweb-meetings.unoda.org/public/2025-12/AIxBio%20Recommendations%20INHR.pdf" rel="">Recommendations to Governments on Mitigating AIxBio Risks</a>” as part of the INHR/CNAS trilateral dialogue.</li>
<li>Participated in roundtables on CBRN (chemical, biological, radiological, and nuclear) risks and on responsible innovation in AI for peace and security, hosted by the Stockholm International Peace Research Institute and the United Nations Office for Disarmament Affairs (UNODA).</li>
<li>Participated in a workshop “<a href="https://ibbis.bio/international-meeting-advances-standards-for-dna-synthesis-screening/" rel="">International Standards for DNA Synthesis Screening – Towards Common Global Standards for Biosecurity</a>” organised by the International Biosecurity and Biosafety Initiative for Science (IBBIS). The meeting marked the launch of the IBBIS <a href="https://ibbis.bio/our-work/international-screening-standards/" rel="">International Standards Initiative</a>.</li>
<li>Spoke on a panel “<a href="https://www.youtube.com/watch?v=gbj5k1UiejY" rel="">AI-Accelerated Biological Risk: Delving into Asia’s Challenges and Emerging Solutions</a>,” organized by AI Safety Asia (AISA)<img loading="lazy" decoding="async" class="alignnone size-full wp-image-1389" src="https://concordia-ai.com/wp-content/uploads/2026/03/image.jpg" alt="" width="1280" height="765" srcset="https://concordia-ai.com/wp-content/uploads/2026/03/image.jpg 1280w, https://concordia-ai.com/wp-content/uploads/2026/03/image-300x179.jpg 300w, https://concordia-ai.com/wp-content/uploads/2026/03/image-1024x612.jpg 1024w, https://concordia-ai.com/wp-content/uploads/2026/03/image-150x90.jpg 150w, https://concordia-ai.com/wp-content/uploads/2026/03/image-768x459.jpg 768w" sizes="auto, (max-width: 1280px) 100vw, 1280px" />
<div class="captioned-image-container">
<figure><figcaption class="image-caption"><span style="color: #999999;"><em>Concordia AI CEO Brian Tse speaking at the United Nations side event during the Sixth Expert Meeting of the Working Group on strengthening implementation of the Biological Weapons Convention (BWC) in Geneva. Source: <a style="color: #999999;" href="https://mp.weixin.qq.com/s/BZ8a4lIQ-b-__-LYiAIfiw" rel="">Concordia AI</a>.</em></span></figcaption></figure>
</div>
<h2 class="header-anchor-post">International research and public engagement</h2>
<ul>
<li><strong>Analysis of China’s AI safety and governance landscape</strong>
<ul>
<li>Published the <em><a href="https://concordia-ai.com/research/state-of-ai-safety-in-china-2025/" rel="">State of AI Safety in China 2025</a></em><a href="https://concordia-ai.com/research/state-of-ai-safety-in-china-2025/" rel=""> report</a>, which was cited by <em><a href="https://www.wired.com/story/china-artificial-intelligence-policy-laws-race" rel="">Wired</a></em>, <em><a href="https://www.bloomberg.com/news/articles/2025-07-30/china-prepares-to-unseat-us-in-fight-for-4-8-trillion-ai-market" rel="">Bloomberg</a></em>, and <em><a href="https://paper.people.com.cn/rmrb/pc/content/202507/31/content_30092070.html" rel="">People’s Daily</a></em>; discussed the findings in a <a href="https://www.youtube.com/watch?v=os2t6vczu00" rel="">webinar</a> with distinguished experts; gave briefings on the report to senior leadership at over ten global organisations.</li>
<li>Our analysis was featured in multiple <em>Nature News</em> stories, including on China’s proposal for a <a href="https://www.nature.com/articles/d41586-025-03902-y" rel="">World Artificial Intelligence Cooperation Organization (WAICO)</a>, on <a href="https://www.nature.com/articles/d41586-025-03845-4" rel="">DeepSeek’s CEO LIANG Wenfeng (梁文锋)</a>, and on <a href="https://www.nature.com/articles/d41586-025-03972-y" rel="">China’s domestic AI governance</a>.</li>
<li>Published 20 “<a href="https://aisafetychina.substack.com/" rel="">AI Safety in China</a>” newsletters, growing our subscriber base by 73% over the course of 2025.</li>
<li>Brian Tse appeared on Nathan Labenz’ <a href="https://www.cognitiverevolution.ai/chinese-ai-they-re-just-like-us-with-beijing-based-concordia-ai-ceo-brian-tse/" rel="">The Cognitive Revolution podcast</a> to discuss China’s approach to AI development, safety, and governance and <em><a href="https://news.cgtn.com/news/2025-03-10/Watch-Youth-driven-growth-in-the-private-economy-1BDcD9ziV4A/p.html" rel="">CGTN</a></em> on China’s approaches in AI innovation and global governance.</li>
<li>Our International AI Governance Senior Research Manager Jason ZHOU (周杰晟) and our International AI Governance Part-time Researcher Gabriel Wagner analysed the AI safety implications of China’s April Politburo study session in a piece for the <a href="https://digichina.stanford.edu/work/forum-xis-message-to-the-politburo-on-ai/" rel="">Stanford DigiChina Forum</a>.</li>
<li>Brian Tse authored an <a href="https://time.com/7308857/china-isnt-ignoring-ai-regulation-the-u-s-shouldnt-either/" rel="">op-ed</a> in <em>Time Magazine</em>, suggesting practical steps for AI safety dialogue between China and the US.</li>
<li>Kwan Yee NG (吴君仪) and Gabriel Wagner spoke on China’s AI safety approach at <a href="https://www.youtube.com/watch?v=SjfnaOEdV80" rel="">AI Safety Asia’s Beijing Roundtable</a>.</li>
</ul>
</li>
</ul>
</li>
</ul>
</li>
</ul>
<p><img loading="lazy" decoding="async" class="alignnone wp-image-710 size-full" src="https://concordia-ai.com/wp-content/uploads/2025/07/CN2-scaled-e1772778001221.jpg" alt="" width="1810" height="1335" srcset="https://concordia-ai.com/wp-content/uploads/2025/07/CN2-scaled-e1772778001221.jpg 1810w, https://concordia-ai.com/wp-content/uploads/2025/07/CN2-scaled-e1772778001221-300x221.jpg 300w, https://concordia-ai.com/wp-content/uploads/2025/07/CN2-scaled-e1772778001221-1024x755.jpg 1024w, https://concordia-ai.com/wp-content/uploads/2025/07/CN2-scaled-e1772778001221-150x111.jpg 150w, https://concordia-ai.com/wp-content/uploads/2025/07/CN2-scaled-e1772778001221-768x566.jpg 768w, https://concordia-ai.com/wp-content/uploads/2025/07/CN2-scaled-e1772778001221-1536x1133.jpg 1536w" sizes="auto, (max-width: 1810px) 100vw, 1810px" /></p>
<ul>
<li><strong>Contributing to international AI safety and governance research</strong>
<ul>
<li>Kwan Yee Ng contributed to the first <em><a href="https://internationalaisafetyreport.org/" rel="">International AI Safety Report</a></em> in 2025 and <a href="https://internationalaisafetyreport.org/publication/international-ai-safety-report-2026" rel="">the second edition in 2026</a> as a writer. The Report provides an up-to-date, internationally shared and science-based understanding of Alcapabilities and risks. It was overseen by an international Expert Advisory Panel nominated by over 30 countries and intergovernmental organisations.</li>
<li>Contributed to the <a href="https://aisafetypriorities.org/" rel="">Singapore Consensus on Global AI Safety Research Priorities</a>, alongside 100 researchers from 11 countries.</li>
<li>Co-published the report <em><a href="https://concordia-ai.com/research/examining-ai-safety-as-a-global-public-good/" rel="">Examining AI Safety as a Global Public Good</a></em> alongside the Carnegie Endowment for International Peace and the Oxford Martin School AI Governance Initiative.</li>
<li>Team members contributed to major papers on autonomous and agentic AI risks: Brian Tse and our AI Safety Research Manager DUAN Yawen (段雅文) contributed to “<a href="https://arxiv.org/abs/2511.22619" rel="">AI Deception: Risks, Dynamics, and Controls</a>” (Peking University), and Duan Yawen worked on the World Economic Forum white paper “<a href="https://www.weforum.org/publications/ai-agents-in-action-foundations-for-evaluation-and-governance/" rel="">AI Agents in Action: Foundations for Evaluation and Governance</a>”, <a href="https://aiagentindex.mit.edu/" rel="">2025 AI Agent Index led by MIT</a>, and the research paper “<a href="https://arxiv.org/abs/2504.15416" rel="">Bare Minimum Mitigations for Autonomous AI Development</a>.”</li>
</ul>
</li>
</ul>
<div class="captioned-image-container">
<figure>
<div class="image2-inset can-restack"><picture><source srcset="https://substackcdn.com/image/fetch/$s_!dgGt!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4314ec26-1a27-4919-a448-ee418c6a5253_720x401.png 424w, https://substackcdn.com/image/fetch/$s_!dgGt!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4314ec26-1a27-4919-a448-ee418c6a5253_720x401.png 848w, https://substackcdn.com/image/fetch/$s_!dgGt!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4314ec26-1a27-4919-a448-ee418c6a5253_720x401.png 1272w, https://substackcdn.com/image/fetch/$s_!dgGt!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4314ec26-1a27-4919-a448-ee418c6a5253_720x401.png 1456w" type="image/webp" sizes="100vw" /></picture></div>
<div><img loading="lazy" decoding="async" class="alignnone size-full wp-image-1265" src="https://concordia-ai.com/wp-content/uploads/2025/11/1.png" alt="" width="941" height="525" srcset="https://concordia-ai.com/wp-content/uploads/2025/11/1.png 941w, https://concordia-ai.com/wp-content/uploads/2025/11/1-300x167.png 300w, https://concordia-ai.com/wp-content/uploads/2025/11/1-150x84.png 150w, https://concordia-ai.com/wp-content/uploads/2025/11/1-768x428.png 768w" sizes="auto, (max-width: 941px) 100vw, 941px" /></div>
<div>
<div class="captioned-image-container">
<figure><figcaption class="image-caption"><span style="color: #999999;"><em>Participants of the 2025 Singapore Conference on AI: International Scientific Exchange on AI Safety. Source: <a style="color: #999999;" href="https://aisafetypriorities.org/" rel="">The Singapore Consensus on Global AI Safety Research Priorities</a>.</em></span></figcaption></figure>
</div>
<ul>
<li><strong>Singapore-related research</strong>
<ul>
<li>Published the <em><a href="https://concordia-ai.com/research/state-of-ai-safety-in-singapore/" rel="">State of AI Safety in Singapore</a></em><a href="https://concordia-ai.com/research/state-of-ai-safety-in-singapore/" rel=""> report</a>, the first comprehensive analysis of Singapore’s AI safety ecosystem, led by our International AI Governance Project Manager Jonathan Lee. He also presented the report at a <a href="https://luma.com/wyereks8" rel="">AI governance in Singapore panel</a> organised by Concordia AI in Singapore, a <a href="https://luma.com/zdgck5eh?tk=KY00WB" rel="">talk</a> organised by the Singapore AI Safety Hub, and at EAGxSingapore.</li>
</ul>
</li>
</ul>
<div class="captioned-image-container">
<figure>
<div class="image2-inset can-restack"><picture><source srcset="https://substackcdn.com/image/fetch/$s_!LQyp!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29f2bf1b-c189-4294-9ba0-a8cb8ea91396_724x710.jpeg 424w, https://substackcdn.com/image/fetch/$s_!LQyp!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29f2bf1b-c189-4294-9ba0-a8cb8ea91396_724x710.jpeg 848w, https://substackcdn.com/image/fetch/$s_!LQyp!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29f2bf1b-c189-4294-9ba0-a8cb8ea91396_724x710.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!LQyp!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29f2bf1b-c189-4294-9ba0-a8cb8ea91396_724x710.jpeg 1456w" type="image/webp" sizes="100vw" /></picture></div>
<div><img loading="lazy" decoding="async" class="alignnone wp-image-707 size-full" src="https://concordia-ai.com/wp-content/uploads/2025/07/SG5-scaled-e1772777919416.jpg" alt="" width="1810" height="1810" srcset="https://concordia-ai.com/wp-content/uploads/2025/07/SG5-scaled-e1772777919416.jpg 1810w, https://concordia-ai.com/wp-content/uploads/2025/07/SG5-scaled-e1772777919416-300x300.jpg 300w, https://concordia-ai.com/wp-content/uploads/2025/07/SG5-scaled-e1772777919416-1024x1024.jpg 1024w, https://concordia-ai.com/wp-content/uploads/2025/07/SG5-scaled-e1772777919416-150x150.jpg 150w, https://concordia-ai.com/wp-content/uploads/2025/07/SG5-scaled-e1772777919416-768x768.jpg 768w, https://concordia-ai.com/wp-content/uploads/2025/07/SG5-scaled-e1772777919416-1536x1536.jpg 1536w" sizes="auto, (max-width: 1810px) 100vw, 1810px" /></div>
<div>
<h2 class="header-anchor-post">Contributing to China’s domestic AI safety and governance landscape</h2>
<ul>
<li><strong>Frontier AI safety risk management and best practices</strong>:
<ul>
<li>Co-published the <em><a href="https://aisafetychina.substack.com/p/shanghai-ai-lab-and-concordia-ai" rel="">Frontier AI Risk Management Framework</a></em><a href="https://aisafetychina.substack.com/p/shanghai-ai-lab-and-concordia-ai" rel=""> v1.0</a> with Shanghai AI Lab. This is China’s first comprehensive framework for managing severe risks from general-purpose AI models.
<ul>
<li>The framework proposes a robust set of protocols designed to support general-purpose AI developers, with comprehensive guidelines for proactively identifying, assessing, mitigating, and governing a set of severe AI risks that pose threats to public safety and national security.</li>
<li>The framework outlines a set of unacceptable hazards (red lines) and early warning indicators for escalating safety and security measures (yellow lines) for areas including: cyber offense, biological threats, large-scale persuasion and harmful manipulation, and loss of control risks.</li>
<li>The framework was cited in various media outlets, including <em><a href="https://science.caixin.com/m/2025-07-30/102346902.html" rel="">Caixin</a></em>, <em><a href="https://mp.weixin.qq.com/s/EwDrlAveGkMm7NsnqCZi6Q" rel="">IT Times</a>,</em> <em><a href="https://www.xinhuanet.com/liangzi/20251112/3687c6b01ecf42ddbd0fd7d9600cc787/c.html" rel="">Xinhua</a></em>, <em><a href="https://time.com/7308857/china-isnt-ignoring-ai-regulation-the-u-s-shouldnt-either/" rel="">TIME</a></em>, <em><a href="https://stock.finance.sina.com.cn/stock/view/paper.php?symbol=sh000001&amp;reportid=807179279125" rel="">Sina</a></em>, and <em><a href="https://www.sinicapodcast.com/p/transcript-the-world-ai-conference" rel="">Sinica Podcast</a></em>.</li>
</ul>
</li>
<li>Signed strategic partnership agreements with several leading Chinese general-purpose AI developers to provide advice on AI safety and risk management best practices.</li>
<li>Provided comprehensive advice on compliance with the EU AI Act and General-Purpose AI Code of Practice to leading Chinese general-purpose AI developers. This work included co-hosting a workshop on “EU Code of Practice &amp; Industry Best Practices: Towards a Global Standard for AI Risk Management, Safety and Security” with SaferAI, the Oxford Martin AI Governance Initiative, and the Safe AI Forum.</li>
<li>Presented on frontier AI risk management during a closed-door workshop at the China AI Industry Alliance’s 15th Plenum Meeting, in the context of its <a href="https://aihub.caict.ac.cn/ai_security_and_safety_commitments" rel="">Disclosure of Practices on the AI Security and Safety Commitments</a>.</li>
<li>Presented on risk management for open-weight frontier models at a <a href="https://mp.weixin.qq.com/s/VGXNHVfo_c5A51yKEdrjvQ" rel="">workshop</a> (“Academic Symposium on AI Industry Development and Legislation”) at Tongji University.</li>
</ul>
</li>
</ul>
<p><img loading="lazy" decoding="async" class="alignnone wp-image-698 size-full" src="https://concordia-ai.com/wp-content/uploads/2025/07/英语封面（无版本号）-scaled-e1772778132808.jpg" alt="" width="1810" height="1080" srcset="https://concordia-ai.com/wp-content/uploads/2025/07/英语封面（无版本号）-scaled-e1772778132808.jpg 1810w, https://concordia-ai.com/wp-content/uploads/2025/07/英语封面（无版本号）-scaled-e1772778132808-300x179.jpg 300w, https://concordia-ai.com/wp-content/uploads/2025/07/英语封面（无版本号）-scaled-e1772778132808-1024x611.jpg 1024w, https://concordia-ai.com/wp-content/uploads/2025/07/英语封面（无版本号）-scaled-e1772778132808-150x90.jpg 150w, https://concordia-ai.com/wp-content/uploads/2025/07/英语封面（无版本号）-scaled-e1772778132808-768x458.jpg 768w, https://concordia-ai.com/wp-content/uploads/2025/07/英语封面（无版本号）-scaled-e1772778132808-1536x917.jpg 1536w" sizes="auto, (max-width: 1810px) 100vw, 1810px" /></p>
<ul>
<li><strong>Frontier AI risk monitoring and evaluation:</strong>
<ul>
<li>Contributed to the “<a href="https://concordia-ai.com/research/frontier-ai-risk-management-framework-in-practice-a-risk-analysis-technical-report/" rel="">Frontier AI Risk Management Framework in Practice: A Risk Analysis Technical Report</a>” led by Shanghai AI Lab. We assessed critical risks from more than 20 frontier LLMs in the following areas: cyber offense, biological and chemical risks, persuasion and manipulation, uncontrolled autonomous AI R&amp;D, strategic deception and scheming, self-replication, and collusion. The report was covered by <a href="https://jack-clark.net/2025/07/28/import-ai-422-llm-bias-china-cares-about-the-same-safety-risks-as-us-ai-persuasion/" rel="">Jack Clark’s </a><em><a href="https://jack-clark.net/2025/07/28/import-ai-422-llm-bias-china-cares-about-the-same-safety-risks-as-us-ai-persuasion/" rel="">Import AI</a></em>.</li>
<li>Launched the <a href="https://airiskmonitor.net:18615/doc/en/report/202507" rel="">AI Risk Monitoring Platform</a> designed to track and mitigate frontier AI risks, including cyberoffense, biological threats, chemical threats, and loss of control domains. The platform evaluates 50 frontier LLMs from 15 leading developers across the US, China, and France, using 18 open source benchmarks. Key outputs include a risk index dashboard and a detailed technical report. This project was spearheaded by our AI Safety Research Senior Manager WANG Weibing (王伟冰).</li>
<li>The platform received coverage from several major media outlets, including <em><a href="https://www.peopleapp.com/column/30050743110-500007199542" rel="">People’s Daily</a></em>, <em><a href="https://www.scmp.com/tech/tech-trends/article/3331952/chinese-ai-models-comparable-us-ones-frontier-risks-study-finds" rel="">South China</a> <a href="https://www.scmp.com/tech/policy/article/3334376/deepseek-alibaba-researchers-endorse-chinas-misunderstood-ai-regulatory-framework" rel="">Morning Post</a></em>, <em>Xinhua</em>’s <em><a href="https://www.jjckb.cn/20251119/dbb052fbebf9490b926eb42d5f4db308/c.html" rel="">Economic Information Daily</a></em>, and <em><a href="https://eu.36kr.com/en/p/3556948076575618" rel="">IT Times</a></em>.</li>
</ul>
</li>
</ul>
<p><img loading="lazy" decoding="async" class="alignnone size-full wp-image-1391" src="https://concordia-ai.com/wp-content/uploads/2026/03/image-scaled.png" alt="" width="2560" height="1303" srcset="https://concordia-ai.com/wp-content/uploads/2026/03/image-scaled.png 2560w, https://concordia-ai.com/wp-content/uploads/2026/03/image-300x153.png 300w, https://concordia-ai.com/wp-content/uploads/2026/03/image-1024x521.png 1024w, https://concordia-ai.com/wp-content/uploads/2026/03/image-150x76.png 150w, https://concordia-ai.com/wp-content/uploads/2026/03/image-768x391.png 768w, https://concordia-ai.com/wp-content/uploads/2026/03/image-1536x782.png 1536w, https://concordia-ai.com/wp-content/uploads/2026/03/image-2048x1042.png 2048w" sizes="auto, (max-width: 2560px) 100vw, 2560px" /></p>
<div>
<div class="single-post-container" role="main" aria-label="Post">
<div class="container">
<div class="single-post">
<article class="typography newsletter-post post">
<div class="available-content">
<div class="body markup" dir="auto">
<ul>
<li><strong>National standards and policy guidance</strong>: Concordia AI is a member of key national and industry technical committees, contributing to the development of China’s AI safety standards.
<ul>
<li>National Information Security Standardization Technical Committee (SAC/TC260): As part of SAC/TC260 Special Working Group on Emerging Technology Safety, Concordia AI contributed to the standard for <strong>“Classification and Grading Methods for the Security of Artificial Intelligence Applications.”</strong></li>
<li>National Information Technology Standardization Technical Committee (SAC/TC28/SC42): As a member of the AI Subcommittee, Concordia AI contributed to <strong>“Artificial intelligence—Risk management capability assessment.”</strong></li>
<li><strong>Ministry of Industry and Information Technology AI Standardization Committee (MIIT/TC1)</strong>: Concordia AI joined the Working Group on AI Safety Governance.</li>
<li>Guangdong-Hong Kong-Macao Greater Bay Area local standards: As a member of the Greater Bay Area working group of SAC/TC28/SC42, Concordia AI played a key role in the development of the Shenzhen local standard<strong> “Technical Framework for Value Alignment of Pre-trained AI Models.”</strong></li>
</ul>
</li>
<li><strong>AIxBiosecurity Governance:</strong>
<ul>
<li>Published a Chinese language report “<a href="https://concordia-ai.com/research/responsible-innovation-in-ai-x-life-sciences/" rel="">Responsible Innovation in AI x Life Sciences</a>” with Tianjin University’s Center for Biosafety Research and Strategy. The 70-page report draws on more than 300 sources to explore AI-biotech convergence, and its benefits, risks, and governance recommendations for diverse stakeholders.
<ul>
<li>Head of AI Safety and Governance (China) FANG Liang (方亮) presented the report at a biosecurity seminar hosted by <a href="https://mp.weixin.qq.com/s?__biz=Mzg4NTgxNjEwMg==&amp;mid=2247501825&amp;idx=1&amp;sn=6f1e0e492aec0f1ae294f2c0ff116f00&amp;scene=21#wechat_redirect" rel="">China’s National Key Laboratory of Synthetic Biotechnology</a>.</li>
<li>Presented the report at the <a href="https://www.linkedin.com/posts/anita-cicero-76900b3a_i-am-pleased-to-have-participated-in-serious-activity-7381972208744550400-0axx/" rel="">2025 International Symposium on Global Biosecurity Governance and Cooperation</a>, co-hosted by the National Biosecurity Expert Committee of China, Guangzhou Laboratory, and China Foreign Affairs University.</li>
<li>Presented the report at the “Symposium on Trends and Development Strategies for the Integration of Biotechnology and AI” hosted by the <a href="https://www.cncb.ac.cn/" rel="">China National Center for Bioinformation</a>.</li>
</ul>
</li>
<li>Participated in the “Closed-door Seminar on DNA Synthesis Screening Technology and Policy” held at China Foreign Affairs University.</li>
</ul>
</li>
<li><strong>WeChat publications:</strong>
<ul>
<li>Published over 80 new posts in our WeChat Official Account, reaching over 4,900 subscribers across China’s AI ecosystem, including policymakers, industry professionals, and academic researchers.</li>
<li>The articles provide Chinese stakeholders with updates on key global AI safety and governance developments. Highlights include a <a href="https://mp.weixin.qq.com/s/7noQvr-ka_JHlYiBRC09OQ" rel="">series of articles</a> on frontier AI safety frameworks by our AI Safety and Governance Senior Manager CHENG Yuan (程远); legal explainers on the <a href="https://mp.weixin.qq.com/s/JEEOVqxKks30pmuJLJStfQ" rel="">EU General-Purpose AI Code of Practice</a> and <a href="https://mp.weixin.qq.com/s/P8G0U8vZZh4DWt0WhliZUA" rel="">California SB-53</a>; and overviews of technical AI safety research on topics like <a href="https://mp.weixin.qq.com/s/AKow85xZYwZwzEPEDfhuOw" rel="">deception</a> and <a href="https://mp.weixin.qq.com/s/3SMZIGAHyiM8NMCVymeoxA" rel="">self-replication</a> risks.</li>
</ul>
</li>
</ul>
<h2 class="header-anchor-post">Organizational updates</h2>
<ul>
<li>We established our Singapore office, welcoming our first full-time staff based in Singapore. Singapore’s status as a global hub and international convenor allowed Concordia AI to convene stakeholders from China, the US, and Southeast Asia, significantly expanding our capacity to shape international discussions on AI governance.</li>
<li>We expanded the team from eight to twelve members in 2025, and will soon grow to 18 staff.</li>
<li>We recruited our seventh cohort of 34 affiliates, who supported both China-focused and international workstreams.</li>
<li>We launched a new <a href="https://concordia-ai.com/" rel="">website</a> with refreshed branding, and updated our brochure and <a href="https://mp.weixin.qq.com/s/7RQ4SL1Vy1PEDtG0gBpwHw" rel="">Chinese materials</a>.</li>
<li>We became a formal member of the <a href="https://partnershiponai.org/partnership-on-ai-welcomes-10-new-partners/" rel="">Partnership on AI</a> and the <a href="https://www.iaseai.org/affiliates" rel="">International Association of Safe and Ethical AI</a> affiliate program.</li>
</ul>
<div class="captioned-image-container" style="text-align: center;">
<figure>
<div class="image2-inset can-restack">
<p><img loading="lazy" decoding="async" class="alignnone size-full wp-image-402" src="https://concordia-ai.com/wp-content/uploads/2025/07/concordia-ai-team-photo-scaled.jpg" alt="" width="2560" height="1281" srcset="https://concordia-ai.com/wp-content/uploads/2025/07/concordia-ai-team-photo-scaled.jpg 2560w, https://concordia-ai.com/wp-content/uploads/2025/07/concordia-ai-team-photo-300x150.jpg 300w, https://concordia-ai.com/wp-content/uploads/2025/07/concordia-ai-team-photo-1024x512.jpg 1024w, https://concordia-ai.com/wp-content/uploads/2025/07/concordia-ai-team-photo-768x384.jpg 768w, https://concordia-ai.com/wp-content/uploads/2025/07/concordia-ai-team-photo-1536x769.jpg 1536w, https://concordia-ai.com/wp-content/uploads/2025/07/concordia-ai-team-photo-2048x1025.jpg 2048w" sizes="auto, (max-width: 2560px) 100vw, 2560px" /></p>
<div class="image-link-expand">
<div class="pencraft pc-display-flex pc-gap-8 pc-reset"><em><span style="color: #999999;">Concordia AI team</span></em></div>
</div>
</div>
</figure>
</div>
<div class="subscription-widget-wrap">
<div class="subscription-widget show-subscribe">
<div class="subscribe-widget is-signed-up is-fully-subscribed" data-component-name="SubscribeWidget">
<div class="pencraft pc-reset button-wrapper">
<div class="pencraft pc-display-flex pc-justifyContent-center pc-reset">
<hr />
</div>
</div>
</div>
</div>
</div>
<div class="footnote" data-component-name="FootnoteToDOM">
<p><a id="footnote-1-189963930" class="footnote-number" contenteditable="false" href="https://aisafetychina.substack.com/p/concordia-ai-2025-impact-highlights?utm_source=post-email-title&amp;publication_id=1862105&amp;post_id=189963930&amp;utm_campaign=email-post-title&amp;isFreemail=true&amp;r=6j32gs&amp;triedRedirect=true&amp;utm_medium=email#footnote-anchor-1-189963930" target="_self" rel="">1</a></p>
<div class="footnote-content">
<p><em>Afternote: We co-organised the “AI Crisis Management” workshop at the sidelines of the Munich Security Conference in February 2026, as a continuation of this workshop series.</em></p>
</div>
</div>
</div>
</div>
</article>
</div>
</div>
</div>
</div>
<div>
<div class="pencraft pc-display-flex pc-flexDirection-column pc-alignItems-center pc-reset color-white-rGgpJs cookieBanner-fZ6hup">
<div class="pencraft pc-display-flex pc-flexDirection-column pc-gap-8 pc-mobile-gap-4 pc-padding-4 pc-alignItems-center pc-reset">
<h4 class="pencraft pc-reset color-primary-zABazT line-height-24-jnGwiv font-display-nhmvtD size-20-P_cSRT weight-bold-DmI9lw reset-IxiVJZ"></h4>
</div>
</div>
</div>
</div>
</figure>
</div>
</div>
</figure>
</div>
<div class="captioned-image-container">
<figure>
<div class="image2-inset can-restack"><picture><source srcset="https://substackcdn.com/image/fetch/$s_!wUhN!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a8a05a9-d0cc-4033-b31e-370f32497b18_1280x765.jpeg 424w, https://substackcdn.com/image/fetch/$s_!wUhN!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a8a05a9-d0cc-4033-b31e-370f32497b18_1280x765.jpeg 848w, https://substackcdn.com/image/fetch/$s_!wUhN!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a8a05a9-d0cc-4033-b31e-370f32497b18_1280x765.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!wUhN!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a8a05a9-d0cc-4033-b31e-370f32497b18_1280x765.jpeg 1456w" type="image/webp" sizes="100vw" /></picture></div>
</figure>
</div><p>The post <a href="https://concordia-ai.com/concordia-ai-2025-impact-highlights/">Concordia AI 2025 Impact Highlights</a> first appeared on <a href="https://concordia-ai.com">Concordia AI</a>.</p>]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>10 Key Insights from Concordia AI’s “Frontier AI Risk Monitoring Platform”</title>
		<link>https://concordia-ai.com/10-key-insights-from-concordia-ais-frontier-ai-risk-monitoring-platform/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=10-key-insights-from-concordia-ais-frontier-ai-risk-monitoring-platform</link>
		
		<dc:creator><![CDATA[Concordia AI]]></dc:creator>
		<pubDate>Tue, 11 Nov 2025 10:52:16 +0000</pubDate>
				<category><![CDATA[Announcement]]></category>
		<guid isPermaLink="false">https://concordia-ai.com/?p=862</guid>

					<description><![CDATA[<p>Concordia AI has launched the Frontier AI Risk Monitoring Platform, along with our inaugural 2025 Q3 Monitoring Report. It tracks models from 15 leading developers worldwide for risks in four domains: cyber offense, biological risks, chemical risks, and loss-of-control, making it the first such platform in China focused on catastrophic risks. You can find more detail, including&#8230;</p>
<p>The post <a href="https://concordia-ai.com/10-key-insights-from-concordia-ais-frontier-ai-risk-monitoring-platform/">10 Key Insights from Concordia AI’s “Frontier AI Risk Monitoring Platform”</a> first appeared on <a href="https://concordia-ai.com">Concordia AI</a>.</p>]]></description>
										<content:encoded><![CDATA[<p>Concordia AI has launched the <a href="https://substack.com/redirect/ec917922-623f-425c-88b5-facbb7840057?j=eyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM" target="_blank" rel="noopener" data-saferedirecturl="https://www.google.com/url?q=https://substack.com/redirect/ec917922-623f-425c-88b5-facbb7840057?j%3DeyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM&amp;source=gmail&amp;ust=1762944030932000&amp;usg=AOvVaw20pQp-Yc3zqm1lPpFFWErC">Frontier AI Risk Monitoring Platform</a>, along with our inaugural <a href="https://substack.com/redirect/3d8361f7-cbe2-4561-b3a1-84d7f8b5638e?j=eyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM" target="_blank" rel="noopener" data-saferedirecturl="https://www.google.com/url?q=https://substack.com/redirect/3d8361f7-cbe2-4561-b3a1-84d7f8b5638e?j%3DeyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM&amp;source=gmail&amp;ust=1762944030932000&amp;usg=AOvVaw0ID9SoQthnHe-nQA-7cwCN">2025 Q3 Monitoring Report</a>. It tracks models from 15 leading developers worldwide for risks in four domains: cyber offense, biological risks, chemical risks, and loss-of-control, making it the first such platform in China focused on catastrophic risks.</p>
<p>You can find more detail, including our <a href="https://substack.com/redirect/6962d40b-4e80-4fd9-b010-c36b839b4b72?j=eyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM" target="_blank" rel="noopener" data-saferedirecturl="https://www.google.com/url?q=https://substack.com/redirect/6962d40b-4e80-4fd9-b010-c36b839b4b72?j%3DeyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM&amp;source=gmail&amp;ust=1762944030932000&amp;usg=AOvVaw3CIW0vO8W5lVZ0bg1d5xyO">methodology</a>, on the interactive platform. The South China Morning Post (SCMP) has also covered the launch in an <a href="https://substack.com/redirect/eecee44e-f9ef-4946-877e-3904fd615999?j=eyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM" target="_blank" rel="noopener" data-saferedirecturl="https://www.google.com/url?q=https://substack.com/redirect/eecee44e-f9ef-4946-877e-3904fd615999?j%3DeyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM&amp;source=gmail&amp;ust=1762944030932000&amp;usg=AOvVaw0OR_2wG97G6HpNVSTNmObt">exclusive story</a>.</p>
<p><img loading="lazy" decoding="async" class="alignnone size-full wp-image-863" src="https://concordia-ai.com/wp-content/uploads/2025/11/Frontier-AI-Risk-Monitoring-Platform.png" alt="" width="1449" height="907" srcset="https://concordia-ai.com/wp-content/uploads/2025/11/Frontier-AI-Risk-Monitoring-Platform.png 1449w, https://concordia-ai.com/wp-content/uploads/2025/11/Frontier-AI-Risk-Monitoring-Platform-300x188.png 300w, https://concordia-ai.com/wp-content/uploads/2025/11/Frontier-AI-Risk-Monitoring-Platform-1024x641.png 1024w, https://concordia-ai.com/wp-content/uploads/2025/11/Frontier-AI-Risk-Monitoring-Platform-150x94.png 150w, https://concordia-ai.com/wp-content/uploads/2025/11/Frontier-AI-Risk-Monitoring-Platform-768x481.png 768w" sizes="auto, (max-width: 1449px) 100vw, 1449px" /></p>
<h2>Why this matters</h2>
<p>As AI capabilities accelerate, we lack insight on some critical questions:</p>
<ul>
<li>What are the key trends and drivers for frontier AI risks?</li>
<li>Are these risks increasing or decreasing?</li>
<li>Where are the safety gaps most severe?</li>
</ul>
<p>Model developers publish self-assessments, but these lack standardization and independent verification. Ad-hoc third-party evaluations don’t track changes over time. Policymakers, researchers, and developers need systematic data to make evidence-based decisions about AI safety. Our platform is our contribution to bridge these gaps.</p>
<h2>10 key insights</h2>
<h4>1. Frontier model risks have risen sharply over the past year</h4>
<p>Across all four domains—cyber offense, biological, chemical, and loss-of-control—Risk Indices for models released in the past year hit record highs. The cumulative maximum Risk Index rose 31% in cyber offense, 38% in biological risks, 17% in chemical risks, and 50% in loss-of-control.</p>
<h4>2. Risk index trends vary significantly across model families</h4>
<p>Over the past year, different model families have followed distinct risk trajectories:</p>
<ul>
<li><strong>Stable low risk: </strong>The GPT and Claude families maintain consistently low Risk Indices across all domains.</li>
<li><strong>Rise then fall: </strong>DeepSeek, Qwen, and MiniMax show early spikes followed by declines in cyber offense, biological, and chemical risks.</li>
<li><strong>Rapid risk increase: </strong>Grok shows sharp increases in loss-of-control risk, while Hunyuan rises steeply in biological risks.</li>
</ul>
<p>Notably, we found that the latest versions of Chinese models released over the past three months have shown a significant decline in risk levels across multiple areas. This is mainly due to stronger refusal of malicious or misuse-related requests.</p>
<h4>3. Reasoning models show higher capabilities without corresponding safety improvements</h4>
<p>Reasoning models score far higher in capability than non-reasoning ones, but their safety levels remain roughly the same. Most models on the Risk Pareto Frontier—a set of models where no other model has both a higher Capability Score and a lower Safety Score—are reasoning models.</p>
<h4>4. The capability and safety performance of open-weight models are generally on par with proprietary models</h4>
<p>The very most capable models are predominantly proprietary, but across the broader landscape, capability and safety levels of open-weight and proprietary models are similar. Only in biological risks do open-weight models score notably lower.</p>
<p><em>Note: Comparable benchmark results do not mean comparable real-world risk. The open-weight nature itself is a key variable affecting risk: it might increase risk by lowering the barrier for malicious fine-tuning; it could also reduce risk by empowering defenders. Due to concerns about misuse, we have set a lower Safety Coefficient for open-weight models, which results in a higher Risk Index compared to proprietary models.</em></p>
<h4>5. Cyberattack capabilities of frontier models are growing rapidly</h4>
<p>Frontier models are showing rapid growth in capabilities across multiple cyberattack benchmarks:</p>
<ul>
<li><strong>WMDP-Cyber (cyberattack knowledge):</strong> Top score rose from 68.9 to 88.0 in one year.</li>
<li><strong>CyberSecEval2-<wbr />VulnerabilityExploit (vulnerability exploitation): </strong>Top score jumped from 55.4 to 91.7.</li>
<li><strong>CyBench (capture the flag): </strong>Top score increased from 25.0 to 40.0.</li>
</ul>
<h4>6. Biological capabilities of frontier models have partially surpassed human expert levels</h4>
<p>Frontier models now match or exceed human experts on several biological benchmarks.</p>
<ul>
<li><strong>BioLP-Bench:</strong> Four models, including o4-mini, outperform human experts in troubleshooting biological protocols.</li>
<li><strong>LAB-Bench-CloningScenarios: </strong>Two models, including Claude Sonnet 4.5 Reasoning, surpass expert performance in cloning experiment scenarios.</li>
<li><strong>LAB-Bench-SeqQA: </strong>The top GPT-5 (high) model nears human-level understanding of DNA and protein sequences (71.5 vs. 79).</li>
</ul>
<h4>7. But most frontier models have inadequate biological safeguards</h4>
<p>Two benchmarks measuring model refusal rates for harmful biological queries show that bio safeguards are lacking:</p>
<ul>
<li><strong>SciKnowEval:</strong> Only 40% of models refused over 80% of harmful prompts, while 35% refused fewer than 50%.</li>
<li><strong>SOSBench-Bio: </strong>Just 15% exceeded an 80% refusal rate, and 35% fell below 20%.</li>
</ul>
<h4>8. Chemical capabilities and safety levels of frontier models are improving slowly</h4>
<p><strong>WMDP-Chem</strong> scores—measuring knowledge relevant to chemical weapons—have risen slightly over the past year, with little variation across models.</p>
<p><strong>SOSBench-Chem</strong> results vary widely: only 30% of models refuse over 80% of harmful queries, while 25% refuse fewer than 40%. Overall, refusal rates show minimal improvement year over year.</p>
<h4>9. Most frontier models have insufficient safeguards against jailbreaking</h4>
<p><strong>StrongReject</strong> evaluates defenses against 31 jailbreak methods. Only 40% of models scored above 80, while 20% fell below 60 (a higher score indicates stronger safeguards). Across all tests, only the Claude and GPT families consistently maintained scores above 80.</p>
<h4>10. Most frontier models fall short on honesty</h4>
<p><strong>MASK</strong> is a benchmark for evaluating model honesty. Only four models scored above 80 points, while 30% of the models scored below 50 points (a higher score indicates a more honest model). Honesty is an important proxy and early warning indicator for loss-of-control risk—dishonest models may misrepresent their capabilities, or provide misleading information about their actions and intentions.</p>
<h2>What’s next</h2>
<p>This is just the beginning. We’re working to:</p>
<ul>
<li>Expand to AI agents, multimodal models, and domain-specific models.</li>
<li>Add new risk domains like large-scale persuasion.</li>
<li>Develop more sophisticated capability elicitation and threat modeling.</li>
<li>Assess both attacker and defender empowerment.</li>
<li>Improve benchmark quality and multilingual coverage.</li>
</ul>
<h2>Get involved</h2>
<p>This is a living project, and we welcome feedback. We’re also seeking partners for benchmark development, risk assessment research, pre-release evaluations, and risk information sharing. More details on avenues for collaboration are available in the <a href="https://substack.com/redirect/a66d29eb-60dd-4328-8e98-97b658c0311e?j=eyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM" target="_blank" rel="noopener" data-saferedirecturl="https://www.google.com/url?q=https://substack.com/redirect/a66d29eb-60dd-4328-8e98-97b658c0311e?j%3DeyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM&amp;source=gmail&amp;ust=1762944030932000&amp;usg=AOvVaw0Kf4RcrBlQlAsLKQ-_yb5f">full report</a>. Contact: <a href="https://substack.com/redirect/8dc7e48c-948f-4fc2-af26-c568b29adb87?j=eyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM" target="_blank" rel="noopener" data-saferedirecturl="https://www.google.com/url?q=https://substack.com/redirect/8dc7e48c-948f-4fc2-af26-c568b29adb87?j%3DeyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM&amp;source=gmail&amp;ust=1762944030932000&amp;usg=AOvVaw1sbKgpRH5qWyL3G76iM43G">risk-monitor@concordia-ai.com</a>.</p><p>The post <a href="https://concordia-ai.com/10-key-insights-from-concordia-ais-frontier-ai-risk-monitoring-platform/">10 Key Insights from Concordia AI’s “Frontier AI Risk Monitoring Platform”</a> first appeared on <a href="https://concordia-ai.com">Concordia AI</a>.</p>]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Concordia AI Joins the Partnership on AI</title>
		<link>https://concordia-ai.com/concordia-ai-joins-the-partnership-on-ai/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=concordia-ai-joins-the-partnership-on-ai</link>
		
		<dc:creator><![CDATA[Concordia AI]]></dc:creator>
		<pubDate>Thu, 23 Oct 2025 07:47:22 +0000</pubDate>
				<category><![CDATA[Announcement]]></category>
		<guid isPermaLink="false">https://concordia-ai.com/?p=854</guid>

					<description><![CDATA[<p>We’re proud to join the Partnership on AI (PAI) in advancing responsible AI development. Together, we aim to promote transparency and accountability, ensuring that AI systems strengthen public trust and serve the common good. Concordia AI has long admired PAI’s work—our CEO Brian Tse previously served as a Senior Advisor to the organization, and we&#8230;</p>
<p>The post <a href="https://concordia-ai.com/concordia-ai-joins-the-partnership-on-ai/">Concordia AI Joins the Partnership on AI</a> first appeared on <a href="https://concordia-ai.com">Concordia AI</a>.</p>]]></description>
										<content:encoded><![CDATA[<p><span style="font-weight: 400;">We’re proud to join the Partnership on AI (PAI) in advancing responsible AI development. Together, we aim to promote transparency and accountability, ensuring that AI systems strengthen public trust and serve the common good.</span></p>
<p><span style="font-weight: 400;">Concordia AI has long admired PAI’s work—our CEO Brian Tse previously served as a Senior Advisor to the organization, and we were thrilled to welcome PAI CEO Rebecca Finlay in Shanghai at the 2025 World AI Conference.</span></p>
<p><span style="font-weight: 400;">We look forward to building on this shared commitment to safe and ethical AI.</span></p>
<p><span style="font-weight: 400;">Learn more about </span><a href="https://partnershiponai.org/partners/?country=china"><span style="font-weight: 400;">PAI’s partners</span></a><span style="font-weight: 400;">. </span></p><p>The post <a href="https://concordia-ai.com/concordia-ai-joins-the-partnership-on-ai/">Concordia AI Joins the Partnership on AI</a> first appeared on <a href="https://concordia-ai.com">Concordia AI</a>.</p>]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Concordia AI Joins the IASEAI Affiliate Program</title>
		<link>https://concordia-ai.com/concordia-ai-joins-the-iaseai-affiliate-program/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=concordia-ai-joins-the-iaseai-affiliate-program</link>
		
		<dc:creator><![CDATA[Concordia AI]]></dc:creator>
		<pubDate>Thu, 23 Oct 2025 07:31:18 +0000</pubDate>
				<category><![CDATA[Announcement]]></category>
		<guid isPermaLink="false">https://concordia-ai.com/?p=851</guid>

					<description><![CDATA[<p>We’re proud to share that Concordia AI has joined the founding cohort of the International Association for Safe and Ethical AI (IASEAI) affiliate program. Our team was honored to participate in the inaugural IASEAI Conference, where Concordia AI CEO Brian Tse spoke on the panel “Global Perspectives on AI Safety and Ethics.” It was an&#8230;</p>
<p>The post <a href="https://concordia-ai.com/concordia-ai-joins-the-iaseai-affiliate-program/">Concordia AI Joins the IASEAI Affiliate Program</a> first appeared on <a href="https://concordia-ai.com">Concordia AI</a>.</p>]]></description>
										<content:encoded><![CDATA[<p><span style="font-weight: 400;">We’re proud to share that Concordia AI has joined the founding cohort of the International Association for Safe and Ethical AI (IASEAI) affiliate program.</span></p>
<p><span style="font-weight: 400;">Our team was honored to participate in the inaugural IASEAI Conference, where Concordia AI CEO Brian Tse spoke on the panel </span><i><span style="font-weight: 400;">“Global Perspectives on AI Safety and Ethics.”</span></i><span style="font-weight: 400;"> It was an inspiring opportunity to exchange ideas on advancing responsible AI governance and safety worldwide.</span></p>
<p><span style="font-weight: 400;">Stay tuned for what’s next—beginning with IASEAI 2026, taking place in Paris this coming February.</span></p>
<p><span style="font-weight: 400;">Learn more about </span><a href="https://www.iaseai.org/affiliates"><span style="font-weight: 400;">IASEAI affiliates</span></a><span style="font-weight: 400;">.</span></p><p>The post <a href="https://concordia-ai.com/concordia-ai-joins-the-iaseai-affiliate-program/">Concordia AI Joins the IASEAI Affiliate Program</a> first appeared on <a href="https://concordia-ai.com">Concordia AI</a>.</p>]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Concordia AI: 2025 Mid-Year Impact Report</title>
		<link>https://concordia-ai.com/concordia-ai-2025-mid-year-impact-report/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=concordia-ai-2025-mid-year-impact-report</link>
		
		<dc:creator><![CDATA[Concordia AI]]></dc:creator>
		<pubDate>Wed, 24 Sep 2025 15:21:30 +0000</pubDate>
				<category><![CDATA[Announcement]]></category>
		<guid isPermaLink="false">https://concordia-ai.com/?p=826</guid>

					<description><![CDATA[<p>Our mission is to ensure that AI is developed and deployed in a way that is safe and aligned with global interests. We advance global AI safety by conducting research, advising leading AI companies and policymakers, and promoting international dialogue. Below are some of our key accomplishments from January to July 2025. (see 2024 highlights&#8230;</p>
<p>The post <a href="https://concordia-ai.com/concordia-ai-2025-mid-year-impact-report/">Concordia AI: 2025 Mid-Year Impact Report</a> first appeared on <a href="https://concordia-ai.com">Concordia AI</a>.</p>]]></description>
										<content:encoded><![CDATA[<p>Our mission is to ensure that AI is developed and deployed in a way that is safe and aligned with global interests. We advance global AI safety by conducting research, advising leading AI companies and policymakers, and promoting international dialogue. Below are some of our key accomplishments from January to July 2025. (<a href="https://substack.com/redirect/94d43be2-5879-4dbd-980d-fc1a34b6abbc?j=eyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM" target="_blank" rel="noopener" data-saferedirecturl="https://www.google.com/url?q=https://substack.com/redirect/94d43be2-5879-4dbd-980d-fc1a34b6abbc?j%3DeyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM&amp;source=gmail&amp;ust=1758798806700000&amp;usg=AOvVaw3bS1KIgqtNWjksWYHRTmOp">see 2024 highlights here</a>).</p>
<h2><strong>Advancing international coordination on AI safety and governance</strong></h2>
<h3><em>Research Impact and Engagement</em></h3>
<ul>
<li><strong>International AI Safety Report:</strong>
<ul>
<li>Head of International AI Governance Kwan Yee Ng (吴君仪) contributed to the <a href="https://substack.com/redirect/fc50b0e5-0c04-423b-bb52-52d323693beb?j=eyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM" target="_blank" rel="noopener" data-saferedirecturl="https://www.google.com/url?q=https://substack.com/redirect/fc50b0e5-0c04-423b-bb52-52d323693beb?j%3DeyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM&amp;source=gmail&amp;ust=1758798806700000&amp;usg=AOvVaw2oH-825vc00RJ7ZF3OnlHB">first International AI Safety Report</a> as one of the writers and is continuing on as a writer for the 2026 edition of the International AI Safety Report. Chaired by Turing Award winner Yoshua Bengio, the report is supported by an expert panel representing 30 countries including China as well as experts from the EU and the UN. Concordia AI also provided feedback to the report, including editing the <a href="https://substack.com/redirect/8c2a4ed6-3412-4b03-9596-c8d360bf0f39?j=eyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM" target="_blank" rel="noopener" data-saferedirecturl="https://www.google.com/url?q=https://substack.com/redirect/8c2a4ed6-3412-4b03-9596-c8d360bf0f39?j%3DeyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM&amp;source=gmail&amp;ust=1758798806700000&amp;usg=AOvVaw3N7nfjwLTb0pnWcMZNCivV">Chinese translation</a> of its summary materials.</li>
</ul>
</li>
<li><strong>China AI safety and governance analysis:</strong>
<ul>
<li>Published the “<a href="https://substack.com/redirect/43009ff2-4b73-4531-a0a4-c828bc2335c7?j=eyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM" target="_blank" rel="noopener" data-saferedirecturl="https://www.google.com/url?q=https://substack.com/redirect/43009ff2-4b73-4531-a0a4-c828bc2335c7?j%3DeyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM&amp;source=gmail&amp;ust=1758798806700000&amp;usg=AOvVaw1j-OrCyrZ6apim2S7lh3ND">State of AI Safety in China 2025</a>” report, covering developments May 2024–June 2025. The report was cited by a number of media outlets including <a href="https://substack.com/redirect/a1a61314-2430-493e-9e48-dc25537d8fc4?j=eyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM" target="_blank" rel="noopener" data-saferedirecturl="https://www.google.com/url?q=https://substack.com/redirect/a1a61314-2430-493e-9e48-dc25537d8fc4?j%3DeyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM&amp;source=gmail&amp;ust=1758798806700000&amp;usg=AOvVaw1FMYui5ORaIdxK4euifGA5">Wired</a>, <a href="https://substack.com/redirect/1de4a0a0-25e6-48cc-9144-1092ca57bd63?j=eyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM" target="_blank" rel="noopener" data-saferedirecturl="https://www.google.com/url?q=https://substack.com/redirect/1de4a0a0-25e6-48cc-9144-1092ca57bd63?j%3DeyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM&amp;source=gmail&amp;ust=1758798806700000&amp;usg=AOvVaw31jTFC4i0uZ518RYt78g9Z">Bloomberg</a>, and <a href="https://substack.com/redirect/0b1b2bca-3537-4408-9bca-a2cc0ff86ca0?j=eyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM" target="_blank" rel="noopener" data-saferedirecturl="https://www.google.com/url?q=https://substack.com/redirect/0b1b2bca-3537-4408-9bca-a2cc0ff86ca0?j%3DeyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM&amp;source=gmail&amp;ust=1758798806700000&amp;usg=AOvVaw1T4Lim67IiSKhRkCzTBIuT">The People’s Daily</a> (the largest newspaper in China).</li>
<li>CEO Brian Tse (谢旻希) authored an op-ed titled “China Is Taking AI Safety Seriously. So Must the U.S.” in <a href="https://substack.com/redirect/1e01c04e-f7af-4460-b8ec-91ea65a06fea?j=eyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM" target="_blank" rel="noopener" data-saferedirecturl="https://www.google.com/url?q=https://substack.com/redirect/1e01c04e-f7af-4460-b8ec-91ea65a06fea?j%3DeyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM&amp;source=gmail&amp;ust=1758798806700000&amp;usg=AOvVaw2dWGQwhtJyvUnNUFHTbjWa">Time Magazine</a>; International AI Governance Senior Research Manager Jason Zhou and International AI Governance Part-time Researcher Gabriel Wagner <a href="https://substack.com/redirect/a1590367-92a2-4f08-86ad-3f47a096a672?j=eyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM" target="_blank" rel="noopener" data-saferedirecturl="https://www.google.com/url?q=https://substack.com/redirect/a1590367-92a2-4f08-86ad-3f47a096a672?j%3DeyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM&amp;source=gmail&amp;ust=1758798806700000&amp;usg=AOvVaw1j2XZmM9ZI7JFAL72YplHI">analyzed</a> AI safety implications of China’s April Politburo study session for Stanford DigiChina Forum; <a href="https://substack.com/redirect/59284c16-f67f-4ab1-958b-a18d2943a877?j=eyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM" target="_blank" rel="noopener" data-saferedirecturl="https://www.google.com/url?q=https://substack.com/redirect/59284c16-f67f-4ab1-958b-a18d2943a877?j%3DeyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM&amp;source=gmail&amp;ust=1758798806700000&amp;usg=AOvVaw2Q4iqESj8ee_Q70bddn_NY">interviewed</a> by CGTN on China’s approaches in AI innovation and global governance.</li>
<li>Published over 10 new &#8220;<a href="https://substack.com/redirect/fcf13d28-bbd9-4d5c-838b-da9c75cb2d1c?j=eyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM" target="_blank" rel="noopener" data-saferedirecturl="https://www.google.com/url?q=https://substack.com/redirect/fcf13d28-bbd9-4d5c-838b-da9c75cb2d1c?j%3DeyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM&amp;source=gmail&amp;ust=1758798806700000&amp;usg=AOvVaw2RplKq2ch3nTYX2-YgJ7oP">AI Safety in China</a>&#8221; newsletter issues, reaching over 1,400 subscribers across governments, top AI labs, and AI safety institutes.</li>
</ul>
</li>
<li><strong>Singapore AI safety and governance analysis:</strong>
<ul>
<li>Published “<a href="https://substack.com/redirect/c17ee9af-02ac-4feb-823f-677b77f55bc1?j=eyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM" target="_blank" rel="noopener" data-saferedirecturl="https://www.google.com/url?q=https://substack.com/redirect/c17ee9af-02ac-4feb-823f-677b77f55bc1?j%3DeyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM&amp;source=gmail&amp;ust=1758798806700000&amp;usg=AOvVaw3mWDuLknLPKqwKBdgOp_Wz">State of AI Safety in Singapore</a>” report, the first comprehensive analysis of Singapore’s AI safety ecosystem.</li>
</ul>
</li>
</ul>
<p><img loading="lazy" decoding="async" class="alignnone wp-image-710" src="https://concordia-ai.com/wp-content/uploads/2025/07/CN2-scaled.jpg" alt="" width="281" height="397" /><img loading="lazy" decoding="async" class="alignnone wp-image-707" src="https://concordia-ai.com/wp-content/uploads/2025/07/SG5-scaled.jpg" alt="" width="281" height="398" /></p>
<h3 style="text-align: left;"><em>Multilateral Initiatives</em></h3>
<ul>
<li><strong>Global AI Summit series: </strong>Participated in the French AI Action Summit, including:
<ul>
<li>Brian Tse was invited as a Chinese civil society representative to the AI Action Summit in the Grand Palais.</li>
<li>Co-hosted the workshop “AI Safety as a Collective Challenge” on the sidelines of the Summit, alongside the Carnegie Endowment for International Peace (CEIP), Oxford Martin AI Governance Initiative (AIGI), Tsinghua University Center for International Security and Strategy (CISS), and Tsinghua Institute for AI International Governance (I-AIIG). During the event, Concordia AI co-published the report “<a href="https://substack.com/redirect/96d2d01b-d7bf-4c88-a81a-90807fc6f388?j=eyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM" target="_blank" rel="noopener" data-saferedirecturl="https://www.google.com/url?q=https://substack.com/redirect/96d2d01b-d7bf-4c88-a81a-90807fc6f388?j%3DeyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM&amp;source=gmail&amp;ust=1758798806700000&amp;usg=AOvVaw11xWgAeAxoO9bQtHNrwAaX">Examining AI Safety as a Global Public Good: Implications, Challenges, and Research Priorities</a>.”</li>
<li>Invited to a closed-door seminar hosted by the China AI Safety &amp; Development Association (CnAISDA). During the subsequent public side event, Turing Award Winner Andrew Yao <a href="https://substack.com/redirect/97efc6a2-3ff3-4a9c-8cfe-5c8280a8ca0a?j=eyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM" target="_blank" rel="noopener" data-saferedirecturl="https://www.google.com/url?q=https://substack.com/redirect/97efc6a2-3ff3-4a9c-8cfe-5c8280a8ca0a?j%3DeyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM&amp;source=gmail&amp;ust=1758798806700000&amp;usg=AOvVaw3P9nliv2E-lSHlgDrmKMlG">cited</a> Concordia AI’s State of AI Safety in China report series when describing the increase in AI safety research in Chinese institutions.</li>
<li>Attended the inaugural conference of International Association of Safe and Ethical AI (IASEAI), where Brian Tse spoke on the panel “Global Perspectives on AI Safety and Ethics.”</li>
<li>Brian Tse delivered a presentation at the France-China AI Association (Association d&#8217;Intelligence Artificielle France-Chine), which was cited by outlets including <a href="https://substack.com/redirect/696b4651-ac60-44fb-b576-86a8529c6591?j=eyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM" target="_blank" rel="noopener" data-saferedirecturl="https://www.google.com/url?q=https://substack.com/redirect/696b4651-ac60-44fb-b576-86a8529c6591?j%3DeyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM&amp;source=gmail&amp;ust=1758798806700000&amp;usg=AOvVaw3hIqruptDFRcPcrTbhJe3t">Xinhua</a>.</li>
</ul>
</li>
<li><strong>United Nations:</strong>
<ul>
<li>Provided <a href="https://substack.com/redirect/c3e3d5de-aab3-4c6e-b476-6df40c89863b?j=eyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM" target="_blank" rel="noopener" data-saferedirecturl="https://www.google.com/url?q=https://substack.com/redirect/c3e3d5de-aab3-4c6e-b476-6df40c89863b?j%3DeyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM&amp;source=gmail&amp;ust=1758798806700000&amp;usg=AOvVaw0A_akokjpOVMuAQAa0cwFh">written inputs</a> and participated in consultations regarding the UN’s Independent International Scientific Panel on AI and Global Dialogue on AI.</li>
<li>Brian Tse <a href="https://substack.com/redirect/7dd6326c-e291-4640-9f43-4d591fdb5f72?j=eyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM" target="_blank" rel="noopener" data-saferedirecturl="https://www.google.com/url?q=https://substack.com/redirect/7dd6326c-e291-4640-9f43-4d591fdb5f72?j%3DeyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM&amp;source=gmail&amp;ust=1758798806700000&amp;usg=AOvVaw1k4BvougF_XVuBJWMr9EBN">spoke</a> on the panel “From Principles to Practice—Governing Advanced AI in Action” at the <a href="https://substack.com/redirect/f3cf7939-9614-409b-a510-26b71fae55b6?j=eyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM" target="_blank" rel="noopener" data-saferedirecturl="https://www.google.com/url?q=https://substack.com/redirect/f3cf7939-9614-409b-a510-26b71fae55b6?j%3DeyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM&amp;source=gmail&amp;ust=1758798806700000&amp;usg=AOvVaw3wOZ4ZXcHlIZdt5cFYVXK5">AI for Good Summit 2025</a>.</li>
</ul>
</li>
<li><strong>Global AIxBiosecurity governance: </strong>We contributed to a number of critical global discussions at the intersection of AI and biosecurity:
<ul>
<li>Brian Tse signed the “<a href="https://substack.com/redirect/77d0cfff-9b28-474c-a17d-15a15eabf3d7?j=eyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM" target="_blank" rel="noopener" data-saferedirecturl="https://www.google.com/url?q=https://substack.com/redirect/77d0cfff-9b28-474c-a17d-15a15eabf3d7?j%3DeyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM&amp;source=gmail&amp;ust=1758798806700000&amp;usg=AOvVaw2Y1RaniI-YflowY_RBNZsN">Statement on Biosecurity Risks at the Convergence of AI and the Life Sciences</a>,” alongside influential world-renowned experts such as Andrew Yao, Yoshua Bengio, and George Church. Provided inputs to the Statement as a member of the <a href="https://substack.com/redirect/6a4c2de6-d5b5-4d98-9d80-97c2b591e520?j=eyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM" target="_blank" rel="noopener" data-saferedirecturl="https://www.google.com/url?q=https://substack.com/redirect/6a4c2de6-d5b5-4d98-9d80-97c2b591e520?j%3DeyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM&amp;source=gmail&amp;ust=1758798806700000&amp;usg=AOvVaw0ef3tlZw3zCS5eNWrYXVPk">AIxBio Global Forum</a>, a platform for international experts and policymakers to identify and reduce biosecurity risks associated with the convergence of AI and the life sciences.</li>
<li>Presented at the &#8220;A Call to Action: The AIxBio Global Forum Statement on Biosecurity Risks at the Convergence of AI and the Life Sciences&#8221; event during <a href="https://substack.com/redirect/3d6aff06-9bd9-4352-a70f-845f57448efe?j=eyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM" target="_blank" rel="noopener" data-saferedirecturl="https://www.google.com/url?q=https://substack.com/redirect/3d6aff06-9bd9-4352-a70f-845f57448efe?j%3DeyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM&amp;source=gmail&amp;ust=1758798806700000&amp;usg=AOvVaw09w5lh79GgPa56EM6-qHyf">The Sixth Session of the Working Group on the Strengthening of the Biological Weapons Convention</a>.</li>
<li>Invited to present at the World Health Organization (WHO) dialogue on the implications of the convergence of AIxBio, for the <a href="https://substack.com/redirect/d5bca258-4928-4a25-8e09-0d11ac96adab?j=eyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM" target="_blank" rel="noopener" data-saferedirecturl="https://www.google.com/url?q=https://substack.com/redirect/d5bca258-4928-4a25-8e09-0d11ac96adab?j%3DeyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM&amp;source=gmail&amp;ust=1758798806700000&amp;usg=AOvVaw1NDsmkwQ71EjHJ5CUM2g_0">Technical Advisory Group on the Responsible Use of the Life Sciences and Dual-Use Research</a>.</li>
<li>Participated in a <a href="https://substack.com/redirect/fe51595b-5c18-4697-9e1c-276a583a3aa1?j=eyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM" target="_blank" rel="noopener" data-saferedirecturl="https://www.google.com/url?q=https://substack.com/redirect/fe51595b-5c18-4697-9e1c-276a583a3aa1?j%3DeyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM&amp;source=gmail&amp;ust=1758798806700000&amp;usg=AOvVaw17SQyzcuIl3UBF-IH9oTjn">AIxBio tabletop exercise at the 2025 Munich Security Conference</a>, hosted by the Nuclear Threat Initiative in collaboration with the Munich Security Conference, which led to the forthcoming report “Safeguarding Against Global Catastrophe: Risks, Opportunities, and Governance Options at the Intersection of AI Intelligence and Biology”.</li>
<li>Participated in a series of roundtable discussions on Responsible Innovation in AI for Peace and Security, including CBRN risks, hosted by Stockholm International Peace Research Institute and the United Nations Office for Disarmament Affairs (UNODA).</li>
<li>Participating in an ongoing track 2 dialogue involving Chinese, American, and international experts, which is formulating policy recommendations for AIxBio governance.</li>
</ul>
</li>
<li><strong>International expert consensus and statements:</strong>
<ul>
<li>Brian Tse participated in the International Dialogues on AI Safety-Shanghai, signing the <a href="https://substack.com/redirect/87b8fdb0-6393-4638-ae8d-70d99c48be13?j=eyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM" target="_blank" rel="noopener" data-saferedirecturl="https://www.google.com/url?q=https://substack.com/redirect/87b8fdb0-6393-4638-ae8d-70d99c48be13?j%3DeyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM&amp;source=gmail&amp;ust=1758798806700000&amp;usg=AOvVaw39M95VUVdIZfpNc9DKsRVT">Shanghai Consensus</a> on Ensuring Alignment and Human Control of Advanced AI Systems, alongside a Nobel laureate, Turing Award winners, and senior policymakers.</li>
<li>Brian Tse and Kwan Yee Ng contributed to and signed <a href="https://substack.com/redirect/b85aae5f-0c85-46b8-91cf-869338594324?j=eyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM" target="_blank" rel="noopener" data-saferedirecturl="https://www.google.com/url?q=https://substack.com/redirect/b85aae5f-0c85-46b8-91cf-869338594324?j%3DeyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM&amp;source=gmail&amp;ust=1758798806700000&amp;usg=AOvVaw3WUGdcKp0vHoGE_ro6CJ5U">The Singapore Consensus on Global AI Safety Research Priorities</a> during the Singapore Conference on AI 2025 (SCAI).</li>
</ul>
</li>
</ul>
<p><img loading="lazy" decoding="async" class="alignnone size-full wp-image-827" src="https://concordia-ai.com/wp-content/uploads/2025/09/unnamed.jpg" alt="" width="1456" height="720" srcset="https://concordia-ai.com/wp-content/uploads/2025/09/unnamed.jpg 1456w, https://concordia-ai.com/wp-content/uploads/2025/09/unnamed-300x148.jpg 300w, https://concordia-ai.com/wp-content/uploads/2025/09/unnamed-1024x506.jpg 1024w, https://concordia-ai.com/wp-content/uploads/2025/09/unnamed-150x74.jpg 150w, https://concordia-ai.com/wp-content/uploads/2025/09/unnamed-768x380.jpg 768w" sizes="auto, (max-width: 1456px) 100vw, 1456px" /></p>
<h2><strong>Convening AI safety conferences in China, Singapore, and globally</strong></h2>
<ul>
<li><strong>World AI Conference Conference (WAIC), Shanghai</strong>
<ul>
<li>Hosted the <a href="https://substack.com/redirect/321328a6-1847-4e28-ad2e-b04356c8a54b?j=eyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM" target="_blank" rel="noopener" data-saferedirecturl="https://www.google.com/url?q=https://substack.com/redirect/321328a6-1847-4e28-ad2e-b04356c8a54b?j%3DeyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM&amp;source=gmail&amp;ust=1758798806700000&amp;usg=AOvVaw1_nGvbsIDEqNEQCPzibAhN">AI Safety and Governance Forum</a> at China&#8217;s most influential AI conference.
<ul>
<li>Convened around 30 distinguished experts, including Yoshua Bengio; United Nations Under-Secretary-General Amandeep Singh Gill; Shanghai AI Lab Director ZHOU Bowen (周伯文); Special Envoy of the President of France for AI Anne Bouverot; Distinguished Professor of Computer Science at UC Berkeley Stuart Russell; Peng Cheng Laboratory Director Academician GAO Wen (高文); CEO of the Partnership on AI Rebecca Finlay; Shanghai Artificial Intelligence Strategic Advisory Expert Committee member Academician HE Jifeng (何积丰); and many more leading figures from government, industry, and research.</li>
<li>Over 200 audience members joined in person, with over 14,000+ views of the livestream, and media coverage from <a href="https://substack.com/redirect/1de4a0a0-25e6-48cc-9144-1092ca57bd63?j=eyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM" target="_blank" rel="noopener" data-saferedirecturl="https://www.google.com/url?q=https://substack.com/redirect/1de4a0a0-25e6-48cc-9144-1092ca57bd63?j%3DeyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM&amp;source=gmail&amp;ust=1758798806700000&amp;usg=AOvVaw31jTFC4i0uZ518RYt78g9Z">Bloomberg</a>, <a href="https://substack.com/redirect/e826da71-4c35-4cd9-8485-8fdb7919a95d?j=eyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM" target="_blank" rel="noopener" data-saferedirecturl="https://www.google.com/url?q=https://substack.com/redirect/e826da71-4c35-4cd9-8485-8fdb7919a95d?j%3DeyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM&amp;source=gmail&amp;ust=1758798806700000&amp;usg=AOvVaw3N_BeTGo-502VzebIpMD09">Wired</a>, <a href="https://substack.com/redirect/9628b07b-d265-4a3f-9e9a-a69c2fea9443?j=eyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM" target="_blank" rel="noopener" data-saferedirecturl="https://www.google.com/url?q=https://substack.com/redirect/9628b07b-d265-4a3f-9e9a-a69c2fea9443?j%3DeyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM&amp;source=gmail&amp;ust=1758798806700000&amp;usg=AOvVaw11C9YfIFwVzUok1qjj8EqM">Caixin</a>, <a href="https://substack.com/redirect/9b762041-c03a-41b8-a2d0-eb4606c9f84c?j=eyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM" target="_blank" rel="noopener" data-saferedirecturl="https://www.google.com/url?q=https://substack.com/redirect/9b762041-c03a-41b8-a2d0-eb4606c9f84c?j%3DeyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM&amp;source=gmail&amp;ust=1758798806700000&amp;usg=AOvVaw2WRDEmMGn4wiv5RXmp3b62">IT Times</a>, and <a href="https://substack.com/redirect/3ae4976b-1b01-442b-8db3-bef35eeb8ca7?j=eyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM" target="_blank" rel="noopener" data-saferedirecturl="https://www.google.com/url?q=https://substack.com/redirect/3ae4976b-1b01-442b-8db3-bef35eeb8ca7?j%3DeyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM&amp;source=gmail&amp;ust=1758798806700000&amp;usg=AOvVaw3HhpIwLLHTWuEzTucRo6IK">Tech Review Africa</a>.</li>
</ul>
</li>
<li>Served as official AI Governance Advisor<strong> </strong>for WAIC 2025.</li>
<li>Co-hosted a number of frontier AI safety workshops on the sidelines of WAIC:
<ul>
<li>Co-hosted a workshop on “Early Warning and Crisis Coordination for Advanced AI” with the Carnegie Endowment for International Peace, Oxford Martin School AI Governance Initiative, Oxford China Policy Lab, Tsinghua University Center for International Security and Strategy (CISS), and Tsinghua University Institute for AI International Governance (I-AIIG).</li>
<li>Co-hosted a workshop on “Convergence of AI and Biological Risks Workshop” with the Tianjin University Center for Biosafety Research.</li>
<li>Hosted a workshop on “Towards International AI Risk Management Standards.”</li>
<li>Co-hosted the “International Workshop on AI Deception Risks and Governance” with Fudan University, Safe AI Forum.</li>
</ul>
</li>
</ul>
</li>
<li><strong>Beijing Academy of AI Conference 2025</strong>
<ul>
<li>Co-hosted the “AI Safety Forum” with the Beijing Academy of Artificial Intelligence (BAAI) at the <a href="https://substack.com/redirect/7915648e-02fd-4c0d-bdbb-9175f033c01e?j=eyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM" target="_blank" rel="noopener" data-saferedirecturl="https://www.google.com/url?q=https://substack.com/redirect/7915648e-02fd-4c0d-bdbb-9175f033c01e?j%3DeyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM&amp;source=gmail&amp;ust=1758798806700000&amp;usg=AOvVaw0BBS6vurQmTOkAxdt_ASAd">BAAI Conference 2025</a>. The forum brought together leading technical experts from institutions including MIT, Fudan University, Singapore Management University, and Tsinghua University to build scientific consensus on technical evaluations for AI “red lines.”</li>
</ul>
</li>
<li><strong>Asia Tech x Singapore, 2025</strong>
<ul>
<li>Organised the <a href="https://substack.com/redirect/62acaa7b-17dc-4f46-b966-dd9a124de3b9?j=eyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM" target="_blank" rel="noopener" data-saferedirecturl="https://www.google.com/url?q=https://substack.com/redirect/62acaa7b-17dc-4f46-b966-dd9a124de3b9?j%3DeyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM&amp;source=gmail&amp;ust=1758798806700000&amp;usg=AOvVaw25COSBT3O62m2k1DPOPM5s">AI Risk Management Workshop</a>, with support from Singapore’s Infocomm Media Development Authority, bringing together 20+ global experts across policy, industry, AI assurance, and academia to explore actionable risk management approaches for AI systems.</li>
</ul>
</li>
<li><strong>International Conference on Learning Representations (ICLR 2025), Singapore</strong>
<ul>
<li>Co-hosted and participated in a series of events, including:
<ul>
<li>Co-hosted the “Frontier Governance Exchange” with Singapore AI Safety Hub, Lorong AI and Safe AI Forum.</li>
<li>Co-convened the “Misalignment and Control Workshop” and a 130+ person AI Safety Social with <a href="http://far.ai/" target="_blank" rel="noopener" data-saferedirecturl="https://www.google.com/url?q=http://FAR.AI&amp;source=gmail&amp;ust=1758798806700000&amp;usg=AOvVaw0Gjd3YCtwZZXfPZQjcB-UT">FAR.AI</a>, the Safe AI Forum, and Singapore AI Safety Hub.</li>
<li>Kwan Yee Ng presented on AI Safety in China at <a href="http://far.ai/" target="_blank" rel="noopener" data-saferedirecturl="https://www.google.com/url?q=http://FAR.AI&amp;source=gmail&amp;ust=1758798806700000&amp;usg=AOvVaw0Gjd3YCtwZZXfPZQjcB-UT">FAR.AI</a>’s <a href="https://substack.com/redirect/e27edb0c-1e5b-433a-b32d-6737026a4713?j=eyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM" target="_blank" rel="noopener" data-saferedirecturl="https://www.google.com/url?q=https://substack.com/redirect/e27edb0c-1e5b-433a-b32d-6737026a4713?j%3DeyJ1IjoiNWVudDAxIn0.zB3sDKg1awnASh1CpBsI3w1DRf4y90PmDO7yXYrrTnM&amp;source=gmail&amp;ust=1758798806700000&amp;usg=AOvVaw3Yq_BUQKBFxWdWfdnT6Fjx">Singapore Alignment Workshop 2025</a>.</li>
</ul>
</li>
</ul>
</li>
</ul>
<p>&nbsp;</p>
<p><img loading="lazy" decoding="async" class="wp-image-751 aligncenter" src="https://concordia-ai.com/wp-content/uploads/2025/09/WAIC-1.png" alt="" width="757" height="505" srcset="https://concordia-ai.com/wp-content/uploads/2025/09/WAIC-1.png 936w, https://concordia-ai.com/wp-content/uploads/2025/09/WAIC-1-300x200.png 300w, https://concordia-ai.com/wp-content/uploads/2025/09/WAIC-1-150x100.png 150w, https://concordia-ai.com/wp-content/uploads/2025/09/WAIC-1-768x512.png 768w" sizes="auto, (max-width: 757px) 100vw, 757px" /></p>
<div><img loading="lazy" decoding="async" class="wp-image-829 aligncenter" src="https://concordia-ai.com/wp-content/uploads/2025/09/9e8ec757-50f7-43f6-8aea-cf7884245f41_6240x4160-scaled.jpg" alt="" width="756" height="504" srcset="https://concordia-ai.com/wp-content/uploads/2025/09/9e8ec757-50f7-43f6-8aea-cf7884245f41_6240x4160-scaled.jpg 2560w, https://concordia-ai.com/wp-content/uploads/2025/09/9e8ec757-50f7-43f6-8aea-cf7884245f41_6240x4160-300x200.jpg 300w, https://concordia-ai.com/wp-content/uploads/2025/09/9e8ec757-50f7-43f6-8aea-cf7884245f41_6240x4160-1024x683.jpg 1024w, https://concordia-ai.com/wp-content/uploads/2025/09/9e8ec757-50f7-43f6-8aea-cf7884245f41_6240x4160-150x100.jpg 150w, https://concordia-ai.com/wp-content/uploads/2025/09/9e8ec757-50f7-43f6-8aea-cf7884245f41_6240x4160-768x512.jpg 768w, https://concordia-ai.com/wp-content/uploads/2025/09/9e8ec757-50f7-43f6-8aea-cf7884245f41_6240x4160-1536x1024.jpg 1536w, https://concordia-ai.com/wp-content/uploads/2025/09/9e8ec757-50f7-43f6-8aea-cf7884245f41_6240x4160-2048x1365.jpg 2048w" sizes="auto, (max-width: 756px) 100vw, 756px" /></div>
<div>
<h2 class="header-anchor-post"><strong>Advising leading AI companies and policymakers in China</strong></h2>
<ul>
<li><strong>National Standards and Policy Guidance: </strong>Concordia AI is a member of key national and industry technical committees, contributing to the development of China’s AI safety standards.
<ul>
<li><strong>National Information Security Standardization Technical Committee (SAC/TC260): </strong>As part of SAC/TC260 Special Working Group on Emerging Technology Safety, Concordia AI contributed to the standard for “Classification and Grading Methods for the Security of Artificial Intelligence Applications.”</li>
<li><strong>National Information Technology Standardization Technical Committee (SAC/TC28/SC42):</strong> As a member of the AI Subcommittee, Concordia AI contributed to the “Artificial intelligence—Risk management capability assessment.”</li>
<li><strong>Ministry of Industry and Information Technology AI Standardization Committee (MIIT/TC1): </strong>Concordia AI joined the Working Group on AI Safety Governance.</li>
<li><strong>Guangdong-Hong Kong-Macao Greater Bay Area local standards: </strong>As a member of the Greater Bay Area working group of SAC/TC28/SC42, Concordia AI played a key role in the development of the Shenzhen local standard “Technical Framework for Value Alignment of Pre-trained AI Models.”</li>
</ul>
</li>
<li><strong>Frontier AI Safety Risk Management and Best Practices:</strong>
<ul>
<li>Co-published the “<a href="https://aisafetychina.substack.com/p/shanghai-ai-lab-and-concordia-ai" rel="">Frontier AI Risk Management Framework v1.0</a>” with Shanghai AI Lab. It is China’s first comprehensive framework for managing severe risks from general-purpose AI models.
<ul>
<li>We propose a robust set of protocols designed to empower general-purpose AI developers, with comprehensive guidelines for proactively identifying, assessing, mitigating, and governing a set of severe AI risks that pose threats to public safety and national security.</li>
<li>The Framework outlines a set of unacceptable outcomes (red lines) and early warning indicators for escalating safety and security measures (yellow lines) for areas including: cyber offense, biological threats, large-scale persuasion and harmful manipulation, and loss of control risks.</li>
</ul>
</li>
<li>Signed strategic partnership agreements with several leading Chinese general-purpose AI developers, providing advice on AI safety and risk management best practices.</li>
<li>Invited to present on frontier AI risk management during a closed-door workshop at the AI Industry Alliance of China’s 15th Plenum Meeting.</li>
<li>Co-hosted a workshop on “EU Code of Practice &amp; Industry Best Practices: Towards a Global Standard for AI Risk Management, Safety and Security” with SaferAI, the Oxford Martin AI Governance Initiative, the Safe AI Forum.</li>
</ul>
</li>
<li><strong>Frontier AI Risk Monitoring and Evaluation:</strong>
<ul>
<li>Contributed to the <a href="https://concordia-ai.com/research/frontier-ai-risk-management-framework-in-practice-a-risk-analysis-technical-report/" rel="">Frontier AI Risk Management Framework in Practice: A Risk Analysis Technical Report</a> led by Shanghai AI Lab. We assessed critical risks from more than 20 frontier LLMs in the following areas: cyber offense, biological and chemical risks, persuasion and manipulation, uncontrolled autonomous AI R&amp;D, strategic deception and scheming, self-replication, and collusion.</li>
<li>Soft-launched an <a href="https://airiskmonitor.net:18615/doc/en/report/202507" rel="">AI Risk Monitoring Platform</a> designed to track and mitigate frontier AI risks, including cyberoffense, biological threats, chemical threats, and loss-of-control domains. The platform evaluates 34 frontier LLMs from 11 leading developers across the U.S., China, and France, using 18 open-source benchmarks. Key outputs include a risk index dashboard and a detailed technical report.</li>
<li>AI Safety Research Manager DUAN Yawen (段雅文) co-authored “<a href="https://arxiv.org/abs/2504.15416" rel="">Bare Minimum Mitigations for Autonomous AI Development</a>.”</li>
</ul>
</li>
<li><strong>AIxBiosecurity Governance:</strong>
<ul>
<li>Published Chinese language report “<a href="https://concordia-ai.com/research/responsible-innovation-in-ai-x-life-sciences/" rel="">Responsible Innovation in AI x Life Sciences</a>” with Tianjin University’s Center for Biosafety Research and Strategy. This 70-page deep dive draws on 300+ sources to explore AI-biotech convergence, benefits, risks, and governance recommendations for diverse stakeholders.
<ul>
<li>Head of AI Safety and Governance (China) FANG Liang (方亮) presented the report at a biosecurity seminar hosted by <a href="https://mp.weixin.qq.com/s?__biz=Mzg4NTgxNjEwMg==&amp;mid=2247501825&amp;idx=1&amp;sn=6f1e0e492aec0f1ae294f2c0ff116f00&amp;scene=21#wechat_redirect" rel="">China’s National Key Laboratory of Synthetic Biotechnology</a>.</li>
<li>Presented the report at the 2025 International Symposium on Global Biosecurity Governance and Cooperation, co-hosted by the National Biosecurity Expert Committee of China, Guangzhou Laboratory, and China Foreign Affairs University.</li>
</ul>
</li>
<li>Invited to participate in the “Closed-door Seminar on DNA Synthesis Screening Technology and Policy” held at China Foreign Affairs University.</li>
</ul>
</li>
<li><strong>WeChat Newsletter Publications:</strong>
<ul>
<li>Released over 59 new posts in our WeChat Official Account, reaching over 4,600 subscribers across China’s AI ecosystem, including policymakers, industry professionals, academic researchers, and the public.</li>
</ul>
</li>
</ul>
<p><img loading="lazy" decoding="async" class="alignnone wp-image-701" src="https://concordia-ai.com/wp-content/uploads/2025/07/双语封面（无版本号）-scaled.jpg" alt="" width="301" height="425" srcset="https://concordia-ai.com/wp-content/uploads/2025/07/双语封面（无版本号）-scaled.jpg 1810w, https://concordia-ai.com/wp-content/uploads/2025/07/双语封面（无版本号）-212x300.jpg 212w, https://concordia-ai.com/wp-content/uploads/2025/07/双语封面（无版本号）-724x1024.jpg 724w, https://concordia-ai.com/wp-content/uploads/2025/07/双语封面（无版本号）-768x1086.jpg 768w, https://concordia-ai.com/wp-content/uploads/2025/07/双语封面（无版本号）-1086x1536.jpg 1086w, https://concordia-ai.com/wp-content/uploads/2025/07/双语封面（无版本号）-1448x2048.jpg 1448w" sizes="auto, (max-width: 301px) 100vw, 301px" /><img loading="lazy" decoding="async" class="alignnone wp-image-681" src="https://concordia-ai.com/wp-content/uploads/2025/07/人工智能-x-生命科学的负责任创新_封面.jpg" alt="" width="301" height="425" srcset="https://concordia-ai.com/wp-content/uploads/2025/07/人工智能-x-生命科学的负责任创新_封面.jpg 596w, https://concordia-ai.com/wp-content/uploads/2025/07/人工智能-x-生命科学的负责任创新_封面-212x300.jpg 212w" sizes="auto, (max-width: 301px) 100vw, 301px" /></p>
<h2 class="header-anchor-post">Organizational updates</h2>
<ul>
<li><strong>Organizational growth:</strong>
<ul>
<li>Following the establishment of our Singapore office, our team expanded from 8 to 12 members, welcoming our first Singapore-based staff member.</li>
</ul>
</li>
<li><strong>International partnerships:</strong>
<ul>
<li>We have strengthened our international engagement by becoming a formal member of the <a href="https://partnershiponai.org/partners/" rel="">Partnership on AI</a> and the <a href="https://www.iaseai.org/" rel="">International Association of Safe and Ethical AI</a>.</li>
</ul>
</li>
<li><strong>Branding and Communication:</strong>
<ul>
<li>We launched a new English organizational <a href="https://concordia-ai.com/" rel="">website</a> with refreshed branding to showcase our work. This is complemented by an updated brochure and a dedicated <a href="https://mp.weixin.qq.com/s/7RQ4SL1Vy1PEDtG0gBpwHw" rel="">WeChat post</a> to provide an introduction to our mission and activities.</li>
</ul>
</li>
</ul>
<p><img loading="lazy" decoding="async" class="alignnone size-full wp-image-402" src="https://concordia-ai.com/wp-content/uploads/2025/07/concordia-ai-team-photo-scaled.jpg" alt="" width="2560" height="1281" srcset="https://concordia-ai.com/wp-content/uploads/2025/07/concordia-ai-team-photo-scaled.jpg 2560w, https://concordia-ai.com/wp-content/uploads/2025/07/concordia-ai-team-photo-300x150.jpg 300w, https://concordia-ai.com/wp-content/uploads/2025/07/concordia-ai-team-photo-1024x512.jpg 1024w, https://concordia-ai.com/wp-content/uploads/2025/07/concordia-ai-team-photo-768x384.jpg 768w, https://concordia-ai.com/wp-content/uploads/2025/07/concordia-ai-team-photo-1536x769.jpg 1536w, https://concordia-ai.com/wp-content/uploads/2025/07/concordia-ai-team-photo-2048x1025.jpg 2048w" sizes="auto, (max-width: 2560px) 100vw, 2560px" /></p>
</div><p>The post <a href="https://concordia-ai.com/concordia-ai-2025-mid-year-impact-report/">Concordia AI: 2025 Mid-Year Impact Report</a> first appeared on <a href="https://concordia-ai.com">Concordia AI</a>.</p>]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Preview: Concordia AI at the 2025 World AI Conference</title>
		<link>https://concordia-ai.com/preview-concordia-ai-at-the-2025-world-ai-conference/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=preview-concordia-ai-at-the-2025-world-ai-conference</link>
		
		<dc:creator><![CDATA[Concordia AI]]></dc:creator>
		<pubDate>Thu, 24 Jul 2025 10:20:40 +0000</pubDate>
				<category><![CDATA[Announcement]]></category>
		<guid isPermaLink="false">https://concordia-ai.com/?p=715</guid>

					<description><![CDATA[<p>We’re excited to share that Concordia AI will be hosting a dedicated AI Safety and Governance Forum at the World AI Conference (WAIC) in Shanghai on July 27, which will be livestreamed globally (full agenda and link below). What is WAIC? WAIC is China’s largest and most influential AI conference, held annually since 2018. It brings together&#8230;</p>
<p>The post <a href="https://concordia-ai.com/preview-concordia-ai-at-the-2025-world-ai-conference/">Preview: Concordia AI at the 2025 World AI Conference</a> first appeared on <a href="https://concordia-ai.com">Concordia AI</a>.</p>]]></description>
										<content:encoded><![CDATA[<p>We’re excited to share that Concordia AI will be hosting a dedicated <strong>AI Safety and Governance Forum</strong> at the World AI Conference (WAIC) in Shanghai on July 27, which will be livestreamed globally (full agenda and link below).</p>
<h2 class="header-anchor-post">What is WAIC?</h2>
<p>WAIC is China’s largest and most influential AI conference, <a href="https://aiii.global/waic-aiii/" rel="">held</a> annually since 2018. It brings together senior Chinese government leaders, foreign diplomats, and renowned experts from across the globe.</p>
<p>WAIC is co-organized by seven central government ministries (MOFA, NDRC, MIIT, MoE, MoST, SASAC, and CAC), along with the Chinese Academy of Sciences (CAS), the China Association for Science and Technology (CAST), and the Shanghai Municipal Government.</p>
<p>In 2024, WAIC was for the first time called a “High-Level Meeting on Global AI Governance”, and saw an unprecedented caliber of senior officials in attendance, most notably Premier LI Qiang (李强). Last year’s edition also increased its focus on AI safety and governance.</p>
<p><a href="https://www.worldaic.com.cn/" rel="">This year’s WAIC</a>, held under the theme “Global Solidarity in the AI Era”, is <a href="https://www.globaltimes.cn/page/202507/1338097.shtml" rel="">expected</a> to break participation records and to maintain its high political profile.</p>
<p>For foreign stakeholders, WAIC is a key event to watch, and a critical opportunity to engage with Chinese counterparts.</p>
<h2 class="header-anchor-post">What to watch at WAIC</h2>
<h3 class="header-anchor-post">Opening ceremony</h3>
<p>The opening ceremony typically draws the highest level of attention, featuring senior Chinese officials, foreign dignitaries, leaders of international organizations, and CEOs of major AI companies.</p>
<p>At WAIC 2024, for instance, Premier Li Qiang, China’s second-highest ranking official, delivered opening remarks emphasizing that AI development must ensure safety, reliability, and controllability. Shanghai Party Secretary CHEN Jining (陈吉宁), one of China’s top 24 officials, also announced the “<a href="https://www.mfa.gov.cn/eng/xw/zyxw/202407/t20240704_11448351.html" rel="">Shanghai Declaration on Global AI Governance</a>,” which warns of “unprecedented challenges” in AI safety and ethics. That same phrase—“unprecedented risks”—was <a href="https://aisafetychina.substack.com/i/162033730/politburo-holds-first-dedicated-ai-development-and-safety-meeting-in-seven-years" rel="">later echoed</a> by President Xi Jinping during a Politburo study session in April 2025, underscoring how WAIC can foreshadow domestic policy trends.</p>
<p>The opening ceremony also features speeches from leading scientists and policy advisors. In 2024, Tsinghua University Institute for AI International Governance (I-AIIG) Dean XUE Lan (薛澜) noted risks of loss of control of AI systems and threats to national security from AI misuse, while Shanghai AI Lab (SHLAB) Director ZHOU Bowen (周伯文) <a href="https://www.shlab.org.cn/news/5443947" rel="">argued</a> that AI safety must be advanced alongside capabilities through alignment, explainability, and reflection in order to achieve trustworthy artificial general intelligence (AGI).</p>
<h3 class="header-anchor-post">Ministerial roundtable</h3>
<p>In 2024, China’s Ministry of Science and Technology (MOST) held a <a href="https://mp.weixin.qq.com/s/id_wBB_McCuBhZm-xvTOeg" rel="">ministerial roundtable</a> with representatives from around 30 other countries. MOST Minister YIN Hejun (阴和俊) noted that AI is an issue “of great importance for the fate of humanity,” while Vice Minister of Foreign Affairs MA Zhaoxu (马朝旭) highlighted the importance of preserving “safety bottom lines” and maintaining human control over AI systems. A similar gathering may occur again this year, even though it remains unconfirmed as of writing.</p>
<h3 class="header-anchor-post">Join us: Concordia AI’s AI Safety and Governance Forum</h3>
<p>WAIC typically features more than 100 thematic “forums,” which often run half-day or day-long. It is impossible to give a full overview here, and we recommend readers check out the full <a href="https://www.worldaic.com.cn/forum" rel="">agenda</a> on the official conference website to get a sense of the range of cutting-edge topics covered. There are multiple forums related to AI safety and governance this year.</p>
<p>We are excited to announce that Concordia AI is hosting the AI Safety and Governance Forum. Last year, <a href="https://aisafetychina.substack.com/p/concordia-ai-holds-the-frontier-ai" rel="">our forum</a> drew almost 300 in-person attendees and over 800,000 livestream views. This year, we’re proud to present another outstanding lineup of leading scientists, policy experts, and industry voices from China and around the world, coming together to discuss frontier AI safety and global governance.</p>
<p>???? Date: July 27<br />
???? Time: 09:00–17:30 (UTC+8)<br />
???? Venue: Shanghai World Expo Center, Meeting Room 518<br />
???? Livestream: <a href="https://online2025.worldaic.com.cn/forumdetail?uuid=F_5oUvbzsk" rel="">Morning session</a> (09:00–12:30 UTC+8); <a href="https://online2025.worldaic.com.cn/forumdetail?uuid=F_0GAIaYvz" rel="">afternoon session</a> (13:35–17:30 UTC+8)</p>
<p>The Forum will be structured around four themes:</p>
<ul>
<li><strong>Theme 1: The Science of AI Safety:</strong> Leading researchers will share insights on the problem “superalignment” and potential pathways to aligned AGI.</li>
<li><strong>Theme 2: Emerging Challenges in AI Safety:</strong> This session examines cutting-edge problems in AI safety, including dual-use challenges in cybersecurity and biosecurity, and early warning signs of deceptive or self-replicating AI systems.</li>
<li><strong>Theme 3: AI Risk Management in Practice:</strong> A comparative look at influential governance approaches across China, the EU, and the US, identifying shared concerns and promising practices for managing AI risk.</li>
<li><strong>Theme 4: International Governance of AI Safety:</strong> The final session explores the evolving landscape of international cooperation on AI safety and potential next steps for global governance.</li>
</ul>
<p><img loading="lazy" decoding="async" class="alignnone wp-image-721" src="https://concordia-ai.com/wp-content/uploads/2025/07/WAIC-2025-agenda-en-scaled.png" alt="" width="516" height="3711" srcset="https://concordia-ai.com/wp-content/uploads/2025/07/WAIC-2025-agenda-en-scaled.png 356w, https://concordia-ai.com/wp-content/uploads/2025/07/WAIC-2025-agenda-en-21x150.png 21w" sizes="auto, (max-width: 516px) 100vw, 516px" /></p><p>The post <a href="https://concordia-ai.com/preview-concordia-ai-at-the-2025-world-ai-conference/">Preview: Concordia AI at the 2025 World AI Conference</a> first appeared on <a href="https://concordia-ai.com">Concordia AI</a>.</p>]]></content:encoded>
					
		
		
			</item>
	</channel>
</rss>
