Winners Consulting Services Co., Ltd. advises executives of Taiwanese companies: In the summer of 2024, Europe finalized two AI regulatory frameworks almost simultaneously—the EU AI Act and the Council of Europe's Framework Convention on Artificial Intelligence, Human Rights, Democracy and the Rule of Law. These two documents have design differences between being "principle-oriented" and using risk-based regulation, which will create a direct compliance divergence for Taiwanese companies exporting to Europe and seeking ISO 42001 certification. Understanding the similarities and differences between these two frameworks is a necessary prerequisite for AI governance planning over the next 3 to 5 years.
Paper Source: Regulating AI from Europe: a joint analysis of the AI Act and the Framework Convention on AI (Miguel Ángel Presno Linera, A. Meuwese, arXiv, 2025)
Original Link: https://doi.org/10.1080/20508840.2025.2492524
About the Authors and This Research
Author Miguel Ángel Presno Linera is a Professor of Constitutional Law at the University of Oviedo in Spain, with a research focus on fundamental rights, digital law, and the EU legal framework. Co-author A. Meuwese has a background in Dutch academia, specializing in comparative regulatory theory and the quality of EU legislation. This paper, published in 2025, has already been cited 8 times as of this writing, making it an early, highly-cited work in the field of comparative AI regulation.
Notably, neither author comes from a purely technical background; they approach AI governance from the perspectives of constitutional law and regulatory theory. This allows their analysis to transcend technical details and address the fundamental institutional tensions between Europe's two AI regulatory instruments—a blind spot often overlooked by Taiwanese companies when planning cross-market AI compliance.
Three Key Comparative Axes of Europe's Dual-Track AI Regulatory Framework
The core contribution of this paper is its systematic comparison of two European AI regulatory documents finalized almost concurrently in the summer of 2024: the EU EU AI Act (applicable to EU member states) and the Council of Europe Framework Convention (with a broader scope covering non-EU signatory countries). The research compares them along three main axes, revealing the institutional gaps that companies must master for practical compliance.
Key Finding 1: Divergent "AI Definitions" Create Ambiguous Application Boundaries
The two documents use different technical and legal standards to define an "artificial intelligence system." The EU AI Act adopts the OECD's definition of AI as its basis, emphasizing machine learning and autonomous reasoning capabilities. The Framework Convention, however, takes a broader, functional definition, focusing on the impact of AI systems on human decision-making. This definitional divergence means that the same AI application might not be classified as a "high-risk AI system" under the EU AI Act but could require stricter fundamental rights scrutiny under the scope of the Framework Convention. For Taiwanese companies exporting to both EU member states and Council of Europe signatories (like the UK, Iceland, Norway), it is essential to assess applicability separately for each framework, as a single compliance document will not suffice.
Key Finding 2: Structural Differences in the Operational Logic of "Risk-Based Regulation"
Although both documents claim to adopt a risk-based regulation approach, their mechanisms for risk classification are fundamentally different. The EU AI Act uses an annex-based list to enumerate high-risk AI application scenarios and sets specific obligations for different risk levels (e.g., transparency requirements, technical documentation, Fundamental Rights Impact Assessment). The Framework Convention, on the other hand, favors principle-based provisions, leaving the discretion for risk assessment to the national legislative bodies of signatory countries. This design difference directly impacts corporate compliance strategies: the EU AI Act's list-based approach provides a relatively clear compliance boundary but risks becoming outdated as technology outpaces legislation. The Framework Convention's principle-based approach requires companies to have stronger autonomous risk assessment capabilities, which aligns well with the dynamic risk management framework required by ISO 42001.
Key Finding 3: The Overall Regulatory Structure Impacts the Global Competitive Landscape
The paper specifically points out that the coexistence of these two European AI regulatory instruments is a strategic move in Europe's competition for global leadership in AI regulation. The EU AI Act represents the EU's "Brussels Effect"—exporting regulatory standards through its single market size. The Framework Convention expands the geographical reach of the European regulatory model through transnational signatures. The synergy or friction between these two mechanisms will directly affect the compliance cost structure for non-European companies entering the European market over the next 3 to 5 years.
Strategic Implications for AI Governance Practices in Taiwan
Taiwanese companies cannot focus solely on the EU AI Act; they must also track the ratification progress of the Council of Europe's Framework Convention and the domestic legislative developments in various countries. This shift in awareness has three specific implications for AI governance planning in Taiwan.
First Layer: Expansion of Compliance Scope. Currently, Taiwanese companies' preparations for EU AI Act compliance are mostly focused on the 27 EU member states. However, if a company's AI products or services are exported to countries that are members of the Council of Europe but not the EU (such as the UK, Norway, Iceland, Turkey), the implementation of the Framework Convention will bring additional Fundamental Rights Impact Assessment obligations. The ISO 42001 management framework provides a common foundation across these frameworks—its risk identification process, required by Clause 6.1.2, can systematically cover the dual assessment needs of the EU AI Act's list-based risks and the Framework Convention's principle-based risks.
Second Layer: Alignment Direction for Taiwan's AI Basic Act. Taiwan's draft AI Basic Act is currently under review in the legislature, and its risk classification structure clearly references the design of the EU AI Act. However, based on this paper's comparative analysis, if Taiwan aims to align with the broader European regulatory system (and not just the EU AI Act), it will need to reserve flexibility for principle-based provisions in its domestic legislation to accommodate the human rights review requirements of the Framework Convention. Taiwanese companies can leverage the dynamic management mechanism of ISO 42001 to build internal governance capabilities that comply with this dual-track framework even before Taiwan's AI Basic Act is finalized.
Third Layer: Product Compliance Impact of AI Definitions. The definitional divergence highlighted in the paper has a direct impact on Taiwanese hardware and software exporters. A product classified as a "general-purpose AI system" under the EU AI Act's definition might be categorized as a high-impact system requiring additional human rights review under the Framework Convention's functional definition. It is recommended that Taiwanese companies incorporate AI risk classification assessments during the product design phase, rather than waiting for market entry to perform compliance remediation.
Winners Consulting Services Helps Taiwanese Companies Build Dual-Track Compliance Capabilities
Winners Consulting Services Co., Ltd. helps Taiwanese companies establish AI management systems that comply with ISO 42001 and the EU AI Act, conduct AI risk classification assessments, ensure that artificial intelligence applications comply with Taiwan's AI Basic Act, and proactively address the human rights impact assessment requirements of the Council of Europe's Framework Convention.
- Dual-Framework AI Definition Mapping and Diagnosis: Based on the definitional differences between the EU AI Act and the Framework Convention revealed in the paper, we conduct a dual-framework applicability assessment of a company's existing AI products and services, clearly identifying high-risk items in the "definitional gray area" to prevent compliance gaps due to misinterpretation.
- Establishment of an ISO 42001 Dynamic Risk Management System: Centered on Clause 6.1.2 of ISO 42001, we design a unified management process that addresses both list-based (EU AI Act) and principle-based (Framework Convention) risk assessments, establishing a quarterly review mechanism to ensure timely updates in line with European regulatory dynamics.
- Internalization of Fundamental Rights Impact Assessment Capabilities: We assist companies in establishing standard operating procedures for Fundamental Rights Impact Assessment in the spirit of the Framework Convention and integrate them into the PDCA management cycle of ISO 42001, preparing them for future compliance audits in the European market.
Winners Consulting Services Co., Ltd. offers a Free AI Governance Mechanism Diagnosis to help Taiwanese companies establish an ISO 42001-compliant management system within 7 to 12 months, while simultaneously addressing the dual-track compliance requirements of the EU AI Act and the European Framework Convention.
Learn More About AI Governance Services → Apply for a Free Mechanism Diagnosis Now →Frequently Asked Questions
- With both the EU AI Act and the Council of Europe's Framework Convention in place, do Taiwanese companies need to prepare two separate sets of compliance documents?
- Not necessarily two completely separate sets of documents, but a management system capable of addressing the differences between the two frameworks is essential. The EU AI Act uses a list-based approach for risk classification, requiring technical documentation, transparency disclosures, and a Fundamental Rights Impact Assessment for high-risk AI systems. In contrast, the Framework Convention relies on principle-based provisions, leaving implementation details to national legislation. The ISO 42001 management framework can serve as a common foundation, as its dynamic risk assessment process systematically covers the core requirements of both. We recommend using ISO 42001 as the central system and adding specific supplementary documents for different markets, rather than building two separate systems from scratch. This approach controls costs while maintaining flexibility for future regulatory changes.
- What are the most common EU AI Act compliance challenges for Taiwanese companies when implementing ISO 42001?
- The three most common challenges are misclassifying AI system risk, structural deficiencies in technical documentation, and mapping compliance between local and EU laws. First, companies often assess risk based on product function rather than the actual application context, underestimating high-risk categories listed in Annex III of the EU AI Act, such as employment screening or credit scoring. Second, while ISO 42001's documentation overlaps with the EU AI Act's requirements, the latter demands more specific details on model descriptions and data governance, which need to be supplemented. Third, companies must create a clear mapping mechanism between Taiwan's AI Basic Act and the EU AI Act to avoid compliance conflicts. Winners Consulting Services recommends addressing these three checkpoints early in the ISO 42001 implementation process.
- What are the core requirements for ISO 42001 certification, and how long does it typically take for Taiwanese companies to implement it?
- ISO 42001, the only global standard for AI management systems, requires establishing an AI policy, defining organizational roles and responsibilities, identifying and assessing AI risks (aligning with the EU AI Act's logic), managing the AI system lifecycle, and ensuring continuous monitoring and improvement. For Taiwanese companies with an existing ISO 27001 or ISO 9001 foundation, implementation and certification can typically be completed in 6 to 9 months. Companies building a management system from scratch should plan for a 9 to 12-month period. The consulting process at Winners Consulting Services involves four stages: gap analysis (1 month), system design (2-3 months), implementation (3-4 months), and certification audit (1-2 months), with full support including document templates and expert consultation.
- How can the costs and expected benefits of implementing ISO 42001 while aligning with the EU AI Act be assessed?
- Implementation costs vary by company size and management maturity. For a mid-sized Taiwanese tech firm (200-500 employees), the total cost for ISO 42001—including consulting, internal resources, and certification fees—typically ranges from NT$800,000 to NT$2,000,000. Incorporating an EU AI Act compliance review adds approximately 15-30% to this cost. In terms of benefits, ISO 42001 certification provides a significant advantage in European market tenders, as some EU procurement rules now require it. A systematic AI risk management framework also reduces incident response costs and legal expenses from compliance disputes, with an expected return on investment of 200% to 350% over 3 to 5 years.
- Why choose Winners Consulting Services for assistance with AI governance issues?
- Winners Consulting Services Co., Ltd. is one of the few consulting firms in Taiwan with expertise in ISO 42001 implementation, EU AI Act analysis, and tracking Taiwan's AI Basic Act. Our team integrates expertise from information security management, legal compliance, and AI technology, employing a "Regulation-Standard-Practice" three-tiered approach. This ensures a company's AI governance framework not only meets international standards like ISO 42001 but is also effectively implemented in practice. We offer a complimentary initial mechanism diagnosis to help companies clarify their compliance priorities and reduce implementation risks before committing significant resources, guiding them to establish a compliant AI management system within 7 to 12 months.
FAQ
- EU AI Act與歐洲議會框架公約同時存在,台灣企業需要分別準備兩套合規文件嗎?
- 不一定需要完全獨立的兩套文件,但必須建立能對應兩套框架差異的管理機制。EU AI Act採清單化風險分級,對高風險AI系統要求技術文件、透明度揭露與基本權利影響評估;框架公約則以原則性條文為主,將執行裁量留給各簽署國國內立法。ISO 42001的管理框架可作為共同基盤——其動態風險評估流程能系統性涵蓋兩套框架的核心要求。建議企業以ISO 42001為主軸建立統一管理系統,再針對特定市場追加對應的補充評估文件,而非從零建立兩套獨立體系,這樣既能控制合規成本,又能保持靈活性以因應後續監管演變。
- 台灣企業導入ISO 42001時,最常遇到哪些EU AI Act合規挑戰?
- 最常見的挑戰有三項。第一是AI系統風險分級誤判:企業往往以產品功能而非實際應用情境評估風險等級,導致EU AI Act附錄三所列高風險類別(如就業篩選、信用評估)被低估。第二是技術文件的結構性缺失:ISO 42001要求的AI管理文件雖與EU AI Act的技術文件有所重疊,但後者對模型說明、資料治理的細節要求更為具體,需要額外補充。第三是台灣AI基本法與EU AI Act的條文對應:台灣企業若同時需要符合國內法規與歐盟要求,必須建立清晰的雙法規映射機制,避免合規矛盾。積穗科研建議在ISO 42001導入初期即納入此三項檢核點。
- ISO 42001認證的核心要求是什麼?台灣企業大約需要多久完成導入?
- ISO 42001是目前全球唯一針對AI管理系統的國際標準,核心要求涵蓋:AI政策制定、組織角色責任分配、AI風險識別與評估(對應EU AI Act風險分級邏輯)、AI系統生命週期管理,以及持續監控與改善機制。已建立ISO 27001或ISO 9001管理基礎的台灣企業,通常可在6至9個月內完成ISO 42001導入並通過驗證;若為首次建立管理系統的企業,則建議預留9至12個月準備期。積穗科研的輔導流程分為四個階段:現況診斷(1個月)、機制設計(2至3個月)、導入實施(3至4個月)、驗證稽核(1至2個月)。
- 導入ISO 42001並同步對應EU AI Act的成本與預期效益如何評估?
- 導入成本因企業規模與現有管理成熟度而異。以中型台灣科技企業(員工200至500人)為例,ISO 42001導入的總成本(含顧問費、內部人力、驗證費用)通常落在新台幣80萬至200萬元之間,若同步納入EU AI Act合規審查,約需額外增加15至30%的評估作業成本。通過ISO 42001認證的企業在歐洲市場的招標資格審查中具有明顯優勢,部分歐盟採購規範已將AI管理系統認證列為供應商資格條件。預期3至5年的投資回收比例可達200%至350%,並可降低AI應用上線後的事故響應與法律合規爭議成本。
- 為什麼找積穗科研協助AI治理相關議題?
- 積穗科研股份有限公司(Winners Consulting Services Co. Ltd.)是台灣少數同時具備ISO 42001輔導能力、EU AI Act法規分析專業與台灣AI基本法政策追蹤能力的顧問機構。顧問團隊整合資訊安全管理、法律合規與AI技術三個專業領域,採用「監管-標準-實務」三層整合方法,確保企業的AI治理框架既符合ISO 42001國際標準,又能有效落地執行。我們提供免費的初步機制診斷服務,協助企業在投入大量資源前先釐清合規優先順序,降低導入風險,並在7至12個月內完成符合國際標準的AI管理系統建立。
Was this article helpful?
Related Services & Further Reading
Risk Glossary
- ▶
Trustworthy AI Assessment List
這是歐盟AI高階專家小組(HLEG)為實踐「可信賴AI倫理準則」所開發的具體評估工具。企業可藉此清單,系統性地檢視其AI系統是否符合七大關鍵要求,從而有效管理合規風險、增強利害關係人信任,並確保技術的穩健性與安全性。
- ▶
semantic interoperability
語意互通性指不同系統間能交換具有無歧異、共享意義的資料之能力。在AI治理與跨國法規遵循情境中,它確保資料在自動化處理與分析時被正確解讀,是企業實現可信賴AI與降低資料誤用風險的基礎。
- ▶
Autonomy over Self-Representation
「自我表述自主權」指個人控制其身份、經歷與願望如何被呈現及詮釋的權利。在AI招聘等自動化決策情境中,此權利確保求職者能直接表達自我,而非僅由演算法片面解讀。對企業而言,尊重此權利是降低歧視性偏誤、遵循個資法規、建立可信賴AI的關鍵風險管理措施。
- ▶
Algorithmic Hiring Assessments
「演算法招募評估」指運用AI模型自動分析求職者數據(如履歷、測驗)以評估其適任性。此技術常用於大規模招募以提升效率,但企業需注意其潛在的歧視偏見與個資法遵風險,確保評估的公平性與透明度。
- ▶
Comparative gap analysis
「比較性差異分析」是一種系統性方法,用於評估組織現況與多個目標標準(如ISO 42001與歐盟AI法案)之間的差距。此方法協助企業在導入AI等新技術時,識別法遵風險、確定改進的優先順序,並制定具體的行動計畫以彌補管理體系的不足。
Want to apply these insights to your enterprise?
Get a Free Assessment