Australia
Voluntary AI Ethics Principles guide responsible AI development in Australia, with potential reforms under consideration.
Artificial intelligence (AI) has made enormous strides in recent years and has increasingly moved into the public consciousness.
Increases in computational power, coupled with advances in machine learning, have fueled the rapid rise of AI. This has brought enormous opportunities, as new AI applications have given rise to new ways of doing business. It has also brought potential risks, from unintended impacts on individuals (e.g., AI errors harming an individual's credit score or public reputation) to the risk of misuse of AI by malicious third parties (e.g., by manipulating AI systems to produce inaccurate or misleading output, or by using AI to create deepfakes).
Governments and regulatory bodies around the world have had to act quickly to try to ensure that their regulatory frameworks do not become obsolete. In addition, international organizations such as the G7, the UN, the Council of Europe and the OECD have responded to this technological shift by issuing their own AI frameworks. But they are all scrambling to stay abreast of technological developments, and already there are signs that emerging efforts to regulate AI will struggle to keep pace. In an effort to introduce some degree of international consensus, the UK government organized the first global AI Safety Summit in November 2023, with the aim of encouraging the safe and responsible development of AI around the world.
Most jurisdictions have sought to strike a balance between encouraging AI innovation and investment, while at the same time attempting to create rules to protect against possible harms. However, jurisdictions around the world have taken substantially different approaches to achieving these goals, which has in turn increased the risk that businesses face from a fragmented and inconsistent AI regulatory environment. Nevertheless, certain trends are becoming clearer at this stage:
Businesses in almost all sectors need to keep a close eye on these developments to ensure that they are aware of the AI regulations and forthcoming trends, in order to identify new opportunities and new potential business risks. But even at this early stage, the inconsistent approaches each jurisdiction has taken to the core questions of how to regulate AI is clear. As a result, it appears that international businesses may face substantially different AI regulatory compliance challenges in different parts of the world. To that end, this AI Tracker is designed to provide businesses with an understanding of the state of play of AI regulations in the core markets in which they operate. It provides analysis of the approach that each jurisdiction has taken to AI regulation and provides helpful commentary on the likely direction of travel.
Because global AI regulations remain in a constant state of flux, this AI Tracker will develop over time, adding updates and new jurisdictions when appropriate. Stay tuned, as we continue to provide insights to help businesses navigate these ever-evolving issues.
Voluntary AI Ethics Principles guide responsible AI development in Australia, with potential reforms under consideration.
The enactment of Brazil's proposed AI Regulation remains uncertain with compliance requirements pending review.
AIDA expected to regulate AI at the federal level in Canada but provincial legislatures have yet to be introduced.
The Interim AI Measures is China's first specific, administrative regulation on the management of generative AI services.
The Council of Europe is developing a new Convention on AI to safeguard human rights, democracy, and the rule of law in the digital space covering governance, accountability and risk assessment.
The successful implementation of the EU AI Act into national law is the primary focus for the Czech Republic, with its National AI Strategy being the main policy document.
The EU introduces the pioneering EU AI Act, aiming to become a global hub for human-centric, trustworthy AI.
France actively participates in international efforts and proposes sector-specific laws.
The G7's AI regulations mandate Member States' compliance with international human rights law and relevant international frameworks.
Germany evaluates AI-specific legislation needs and actively engages in international initiatives.
National frameworks inform India’s approach to AI regulation, with sector-specific initiatives in finance and health sectors.
Israel promotes responsible AI innovation through policy and sector-specific guidelines to address core issues and ethical principles.
Japan adopts a soft law approach to AI governance but lawmakers advance proposal for a hard law approach for certain harms.
Kenya's National AI Strategy and Code of Practice expected to set foundation of AI regulation once finalized.
Nigeria's draft National AI Policy underway and will pave the way for a comprehensive national AI strategy.
Position paper informs Norwegian approach to AI, with sector-specific legislative amendments to regulate developments in AI.
The OECD's AI recommendations encourage Member States to uphold principles of trustworthy AI.
Saudi Arabia is yet to enact AI Regulations, relying on guidelines to establish practice standards and general principles.
Singapore's AI frameworks guide AI ethical and governance principles, with existing sector-specific regulations addressing AI risks.
South Africa is yet to announce any AI regulation proposals but is in the process of obtaining inputs for a draft National AI plan.
South Korea's AI Act to act as a consolidated body of law governing AI once approved by the National Assembly.
Spain creates Europe's first AI supervisory agency and actively participates in EU AI Act negotiations.
Switzerland's National AI Strategy sets out guidelines for the use of AI, and aims to finalize an AI regulatory proposal in 2025.
Draft laws and guidelines are under consideration in Taiwan, with sector-specific initiatives already in place.
Turkey has published multiple guidelines on the use of AI in various sectors, with a bill for AI regulation now in the legislative process.
Mainland UAE has published an array of decrees and guidelines regarding regulation of AI, while the ADGM and DIFC free zones each rely on amendments to existing data protection laws to regulate AI.
The UK prioritizes a flexible framework over comprehensive regulation and emphasizes sector-specific laws.
The UN's new draft resolution on AI encourages Member States to implement national regulatory and governance approaches for a global consensus on safe, secure and trustworthy AI systems.
The US relies on existing federal laws and guidelines to regulate AI but aims to introduce AI legislation and a federal regulation authority.
Switzerland's National AI Strategy sets out guidelines for the use of AI, and aims to finalize an AI regulatory proposal in 2025.
Currently, there are no specific laws, statutory rules, or regulations in Switzerland that directly regulate AI. AI in Switzerland is currently subject to “traditional” laws that are mostly technology-neutral, such as data protection legislation or medical device regulations, and other sector-specific frameworks.
As noted above, there are currently no specific laws or regulations in Switzerland that directly regulate AI. The Federal Council has instructed the Federal Department of the Environment, Transport, Energy and Communications (DETEC) to identify potential approaches to regulating AI by the end of 2024, with the aim of issuing a mandate for an AI regulatory proposal in 2025.1
Although Switzerland has usually favored a technology-neutral approach with a focus on sectoral rules instead of a horizontal regulation, a report on “Artificial intelligence and international rules” in 2022 found that this approach may cause tensions with the international regulatory framework.2 In any event, it is expected that any future AI regulation in Switzerland will be compatible with international standards to ensure that regulatory rules are not fragmented.
In this context, Switzerland also takes an active role in shaping global AI regulations, such as through discussions within the Council of Europe Committee on Artificial Intelligence.
There are various laws that do not directly seek to regulate AI but may affect the development or use of AI in Switzerland. A non-exhaustive list includes:
In addition, there is soft law and guidance from authorities, including guidelines about artificial intelligence for the federal government (2020)16 recommendations from the Swiss Federal Data Protection and Information Commissioner about data processing in relation to AI (2023)17 and expectations from the Swiss Financial Market Supervisory Authority (FINMA) for the use of AI by regulated institutions (FINMA Risk Monitor 2023)18 (see "Sectoral scope" section below for more examples).
As noted above, there are currently no specific laws or policies in Switzerland that directly regulate AI. Accordingly, no definition of AI is currently recognized through Swiss national legislation. As Swiss laws have focused on a technology-neutral regulation, there has been no need for a distinct legal definition.
While AI has been defined in publications by different government bodies (such as the Federal Department of Foreign Affairs’ “Artificial Intelligence and international rules” report of April 202219, or the Federal Council’s “Challenges of Artificial Intelligence” report of December 201920), such definitions have been inconsistent, and it is not possible to draw firm conclusions about future regulation at the time of writing.
As noted above, there are currently no specific laws or regulations in Switzerland that directly regulate AI. Accordingly, there is no specific territorial scope at this stage. However, it should be noted that existing laws, such as the FADP, may apply extraterritorially under conflict-of-laws rules.
As noted above, there are currently no specific laws or regulations in Switzerland that directly regulate AI. Accordingly, there is no specific sectoral scope at this stage. However:
As noted above, there are currently no specific laws or regulations in Switzerland that directly regulate AI. Accordingly, there are currently no specific or unique obligations imposed on developers, users, operators and/or deployers of AI systems. There are, however, general requirements for the governance in companies, in particular stock companies, which continue to apply in relation to AI, and additional requirements apply for listed entities.
As noted above, there are currently no specific laws or regulations in Switzerland that directly regulate AI. Where authorities have expressed expectations, these focus on robustness and reliability, non-discrimination, transparency, governance and accountability.
As noted above, there are currently no specific laws or regulations in Switzerland that directly regulate AI. Although the EU’s risk categorization is mentioned in the “Artificial intelligence and international rules” report by the Federal Council, the report states that Swiss regulation may not be confined to such an approach.24
As noted above, there are currently no specific laws or regulations in Switzerland that directly regulate AI. However, the existing legal framework must be observed.
As mentioned above, the Federal Council has begun a project to identify a regulatory approach to AI by 2024.25 Such regulation would be subject to the standard legislative procedure. It is unclear whether such regulation would involve a supervisory or regulatory authority, or which authority would exercise any enforcement powers.
However, as indicated above, certain authorities have supervisory and/or regulatory powers under existing laws that do not directly regulate AI but extend to its use in regulated sectors. For example, the Federal Data Protection and Information Commissioner (FDPIC) is the privacy authority and supervises federal data protection regulations. In this capacity, the FDPIC also supervises data protection in the context of AI. Another example is FINMA, which (as noted above) has issued circulars that may apply to the use of AI (although AI is not explicitly mentioned) and has expectations in relation to the use of AI.26 Moreover, FINMA monitors the use of AI in the financial sector and may potentially step in if it determines a regulatory need (although no concrete plans are known).27
In addition, other federal and cantonal governmental bodies address AI and participate in the regulatory processes, for example:
As noted above, there are currently no specific laws or regulations in Switzerland that directly regulate AI. Accordingly, it is currently unclear what enforcement powers the relevant authority will have or what penalties they may impose upon breaches.
As AI is currently governed by “traditional” laws, the regulators may have enforcement powers in relation to breaches of these laws. For instance, the FDPIC may issue binding orders to modify, suspend or discontinue data processing, or to delete personal data (but has no power to impose penalties). Some data protection violations may also be subject to criminal prosecution. In the area of financial market legislation, the FINMA has extensive enforcement tools, including ordering action to restore compliance with the law, industry bans, cease-and-desist orders and activity bans, the disgorgement of profits, withdrawal of authorization, or liquidation of supervised institutions.
1 See here.
2 See here, pages 20 et seq.
3 See the FADP here, article 21.
4 See the Copyright Act here.
5 See the Patents Act here.
6 See the Swiss Civil Code here.
7 See the Code of Obligations here.
8 See the Product Liability Act here.
9 See the Product Safety Act here.
10 See the Therapeutic Products Act here.
11 See the Medical Devices Ordinance here.
12 See the Gender Equality Act here.
13 See the Disability Discrimination Act here.
14 See the Swiss Criminal Code here.
15 See the Federal Constitution here.
16 See here.
17 See here.
18 See here.
19 See here, page 5.
20 See here, page 5.
21 See here.
22 See here.
23 See the Road Traffic Act here, Article 106(5).
24 See here, pages 14 et seq. and page 20.
25 See here.
26 See here.
27 See here.
28 See here.
29 See here.
30 See here.
31 See here.
White & Case means the international legal practice comprising White & Case LLP, a New York State registered limited liability partnership, White & Case LLP, a limited liability partnership incorporated under English law and all other affiliated partnerships, companies and entities.
This article is prepared for the general information of interested persons. It is not, and does not attempt to be, comprehensive in nature. Due to the general nature of its content, it should not be regarded as legal advice.
© 2024 White & Case LLP
David Vasella
Partner, Walder Wyss
+41 58 658 52 87
david.vasella@walderwyss.com
Kento Reutimann
Attorney at Law, Walder Wyss
+41 58 658 55 11
kento.reutimann@walderwyss.com