Contributed By Homburger
At present, there is no legislation in Switzerland dealing specifically with AI, and there is no current initiative to enact general-purpose AI regulation (see 3.1 General Approach to AI-Specific Legislation for further details).
However, the majority of Swiss law is technology-neutral, and can therefore also apply to AI-based products and services. This includes the following.
Notably, Switzerland is neither a member state of the European Union nor a member of the European Economic Area, so EU/EEA legislation only applies to the country where it is specifically referenced in Swiss legislation.
Before the launch of ChatGPT in November 2022, the use of AI in Switzerland focused primarily on predictive AI. Much innovation in AI was concentrated in industries that have historically been strong users of technology in the country, such as finance, pharmaceuticals and medical devices, and robotics. By way of example, financial institutions have been using predictive AI extensively for fraud and money laundering prevention, portfolio and suitability analysis, and trading strategies. The most widespread use of “generative” AI in all industries was arguably text translation, with tools such as DeepL.
Following the general availability of ChatGPT, generative AI tools have seen a massive uptake in various industries in Switzerland. This includes “high impact” industries such as software and technology, media, pharmaceuticals, and finance, as well as industries where the use of generative AI is still in earlier stages, such as marketing and professional services. According to a study conducted by PwC’s strategy consulting “Strategy&” in March 2024, the Swiss economy has the highest growth potential worldwide through generative AI and could expand by 0.5 to 0.8% annually in the coming years based solely on the use of generative AI.
In its “Guidelines on Artificial Intelligence for the Confederation” published on 25 November 2020 (AI Guidelines), the Federal Council (Switzerland’s highest executive authority) makes the case that Switzerland should create and maintain an environment conducive to the research, development and application of AI in order to ensure high-quality living standards. The Swiss government does not specifically facilitate investments in the AI sector or specific technology, but favours a bottom-up approach. There are two essential pillars to incentivise the development and use of AI, as follows.
Consistent with the principle of technology-neutral regulation, the AI Guidelines do not distinguish between different kinds of AI, such as generative AI or predictive AI.
Recently, multiple Swiss public research universities have established dedicated research centres and hubs for AI, aiming to combine researchers from different faculties and industry stakeholders and to facilitate AI start-ups and spin-offs. One prominent example is the ETH AI Center of the Swiss Federal Institute of Technology in Zurich (ETH Zurich).
The Swiss government has historically adopted a “wait-and-see” approach with respect to regulating AI, and has favoured industry self-regulation. However, in light of other regulatory initiatives regarding AI, notably the European Union’s AI Act and the Council of Europe’s proposed AI Convention, the Federal Council decided in November 2023 to review possible regulatory approaches that would be compatible with those two regulatory frameworks.
The results of this review are expected to be available by the end of 2024, and are intended to serve as a basis for the preparation of a specific regulatory proposal. The proposed regulation will then likely undergo an extensive consultation process prior to being introduced to the Swiss parliament.
There is no applicable information in this jurisdiction.
AI Guidelines
The AI Guidelines set out seven basic guidelines on the use of AI in Switzerland. They are not binding on the private sector as their primary purpose is to serve as a general frame of reference for the use of AI within the federal administration and to ensure a coherent policy. They can be summarised as follows.
Fact Sheet Generative AI
In response to the rapid rise of generative AI chatbots, particularly ChatGPT, the Swiss government issued a “Fact sheet on the use of generative AI tools in the Federal Administration”, the most recent version dating from 18 January 2024. The goal of this fact sheet is to give practical guidance to the employees of the federal administration and other federal agencies on how to use generative AI tools for their daily work.
The fact sheet encourages responsible experimentation with generative AI tools, such as summarising publicly available sources, obtaining code suggestions or generating images for presentations. It also reminds users not to violate existing regulations and policies. In particular:
Similar fact sheets for the use of generative AI tools by government employees have been issued at cantonal (state) and local levels, eg, in the Canton of St. Gallen.
There is no applicable information in this jurisdiction.
There is no applicable information in this jurisdiction.
There is no applicable information in this jurisdiction.
There is no applicable information in this jurisdiction.
There is no applicable information in this jurisdiction.
At present, no judicial decisions dealing with AI in substance have been reported in Switzerland.
Given the lack of reported judicial decisions on AI, there are no established definitions of AI used by the courts. In the absence of definitions established by case law, Swiss courts are likely to refer to definitions established by future statutes (if any) or definitions used by the Swiss government, such as the Federal Council’s dispatch accompanying a draft act or guidelines and fact sheets adopted by the federal government or its agencies (see 5.2 Technology Definitions for further information).
OFCOM
The Federal Office of Communications (OFCOM) is considered the leading agency dealing with the regulation of AI in Switzerland. OFCOM’s regulatory authority generally comprises telecommunications, radio and television, and postal services. OFCOM plays an important role in initiatives regarding the digitalisation of the federal administration and is also the office tasked with leading the evaluation of possible regulatory approaches for AI (see 3.1 General Approach to AI-Specific Legislation).
FDPIC
The Federal Data Protection and Information Commissioner (FDPIC) is an independent agency responsible for tasks in the areas of data protection and the principle of freedom of information. In its role as supervisory authority, the FDPIC monitors and enforces compliance by federal bodies and the private sector with Swiss federal data protection legislation. The FDPIC’s views and guidelines on the application of data protection legislation to AI are of significant practical importance.
FINMA
FINMA is the independent agency responsible for financial-market supervision in Switzerland. FINMA’s regulatory mandate is to supervise banks, insurance companies, exchanges and financial institutions, among others. It is charged with protecting creditors, investors and policyholders and ensuring that Switzerland’s financial markets function effectively. While FINMA’s regulatory authority is not specifically aimed at AI, it has recently identified AI as a key trend in its risk monitoring published in November 2023. FINMA stated that AI creates particular challenges in the financial sector in connection with the responsibility for AI decisions, the reliability of AI applications, the transparency and “explainability” of AI decisions, and the equal treatment of financial market clients. In its report, FINMA also states that it will review the use of AI by supervised institutions in line with the risk-based approach and the principle of proportionality.
CNAI
The Swiss federal government’s Competence Network for Artificial Intelligence (CNAI) is not a regulatory agency but is the only AI-specific body established within the federal government. The CNAI combines representatives of multiple federal departments and offices, and collaborates with AI experts in its “community of experts” and with other interested individuals, institutions, organisations and companies in its “community of practice”. In addition to drawing on expert knowledge and ensuring that knowledge gained from AI projects is shared within the federal administration, the CNAI has issued an official document for uniform AI terminology in the federal administration (see 5.2 Technology Definitions).
The CNAI has issued an official AI terminology document to introduce uniform use of terms throughout the federal administration. The CNAI terminology includes inter alia the following definitions.
At present, there are no definitions of equivalent significance issued by other government entities that would conflict with the definition of the CNAI.
Please see 5.1 Regulatory Agencies.
So far, there have been no reported enforcement actions regarding the use of AI in Switzerland.
The Swiss Association for Standardization (SNV), a private body and Switzerland’s member of CEN and ISO, is the primary forum for industry standardisation in the country. The SNV has not yet issued any standards related to AI, but is involved in the standards being adopted and considered by ISO.
Most standards in Switzerland are based on the relevant international standards developed in global or European standard-setting bodies. The most important international standard-setting bodies for Switzerland with respect to technology include ISO, IEC, ITU, CEN, CENELEC, and ETSI.
Given that Switzerland does not have AI-specific regulation at present, international standards relating to AI do not have the force of law in Switzerland. However, international standards could be taken into account by the courts when assessing the required standard of care in questions of liability.
Government use of AI in Switzerland varies significantly between the federal government and the different cantonal (state) and local governments, and is still in its early stages overall. While the federal government and most Cantons are already using or evaluating some form of AI, many of those proposed applications do not raise major concerns. Examples include semantic search in legislative texts, summarisation of administrative guidance, automatically assigning incoming questions or requests to the responsible departments for a response, and government chatbots.
Proposed new applications of AI in the areas of policing and criminal justice, on the other hand, have proven more controversial. One example is the Swiss Federal Office of Police (fedpol) working on an upgrade of its automated fingerprint and palmprint identification system (AFIS) to enable facial image comparison. The new system is expected to be operational in 2026, and will provide federal and cantonal immigration authorities, prosecutors, and border security with an additional means for identifying people. Currently, the proposed uses do not include real-time identification of individuals, which is highly controversial.
The most comprehensive list of government use of AI in Switzerland is compiled by the NGO AlgorithmWatch CH in its “Atlas of Automation Switzerland”.
There is no applicable information in this jurisdiction.
To the extent that such information is public, AI does not yet play a key role in national security matters in Switzerland. However, the Swiss Federal Intelligence Service (FIS) has stated in its most recent annual situation report “Switzerland’s Security 2023” that it expects foreign intelligence services to increase their use of AI to process and analyse the ever-increasing amounts of data being transmitted. The FIS also expects intelligence services to further improve their data-gathering capabilities domestically and abroad, which has implications for the FIS’ counter-intelligence and counter-espionage activities.
See 8.2 IP and Generative AI and 8.3 Data Protection and Generative AI.
IP protection of training data and AI models
The degree of IP protection of training data will depend on the nature of the data: if the training data consist of works that are subject to copyright, such as works of literature, newspaper articles, website contents, images, songs or software code, the Copyright Act prohibits unauthorised use, copying or distribution (among other actions). There are statutory limitations to copyright, but many of them will likely not apply in a commercial context (see IP infringement when training generative AI below). The unauthorised copying and distribution under the Copyright Act can trigger civil liability (injunctive relief, damages, disgorgement of profits) and, if committed with intent, constitutes a criminal offence.
Where the training data is not subject to copyright, as would likely be the case for statistical data, measurements, etc., or where copyright protection has expired, the data would have to be protected as such. The same applies to AI models, which, at their core, are simply structured sets of data points (weights, biases, other parameters). With respect to these kinds of data, two main avenues of protection are available, as follows.
IP protection of generative AI output
According to the prevailing doctrine in Switzerland, inventions generated by generative AI are not eligible for patent protection and works created by AI, including texts, images, audio, video and software, are not eligible for copyright protection (see 15.1 Applicability of Patent and Copyright Law).
The output itself may however be protected contractually or as a trade secret if it is kept confidential (see IP protection of training data and AI models above and 15.2 Applicability of Trade Secrecy and Similar Protection).
IP infringement when training generative AI
The training of AI models usually occurs in two steps: the compilation of a training data set and the actual training of the model by iterating over the training data and adjusting the parameters. Both activities involve the creation of copies of data, and can therefore constitute copyright infringement. The Copyright Act does provide for two relevant limitations, but these are unlikely to apply in a commercial context, as follows.
In addition, where the training of an AI model involves the use of unlawfully obtained data that is subject to trade secret protection, the training would also constitute unfair competition under the Unfair Competition Act.
IP infringement by generative AI output
The disclosure or distribution of AI-generated output can infringe on copyright where it reproduces sufficiently large parts of a copyrighted work without authorisation of the copyright holder.
Under the Swiss Federal Act on Data Protection (FADP), the processing of personal data is generally lawful if the principles set out in Articles 6 (eg, proportionality, purpose limitation) and 8 (data security) of the FADP are complied with. A legal justification (consent, overriding private or public interest, or as provided by law) is required for the processing to remain lawful if the processing deviates from these principles, or if the data subject has objected to the processing. This concept differs slightly from the EU GDPR, where a lawful basis is required for all processing.
In the context of generative AI, data protection is relevant both with respect to the training of the AI model and with respect to its operation, particularly regarding the data subject’s right to rectification and deletion, as follows.
Use of AI tools
Apart from translation tools, early applications of AI in the legal profession in Switzerland were primarily focused on reviewing large numbers of documents as part of legal due diligence in M&A, or as part of internal investigations.
However, the availability and use of AI tools is currently undergoing rapid change. Ever since ChatGPT was released to the public, generative AI based on large language models has attracted significant interest from law firms and in-house legal departments. Tasks that Swiss law firms are experimenting with include summarisation, generation of analyses and memoranda based on legal source documents, drafting of contracts, the analysis of weaknesses and inconsistencies in argumentation, etc. Another upcoming application is semantic document search in a law firm’s or legal department’s own document database.
Certain legal information providers are also experimenting with generative AI to offer semantic document search and the AI-based answering of questions based on their database of legal literature, scholarly articles and court cases.
Legal and ethical concerns
The main legal and ethical concerns in the use of AI, and generative AI, in particular, relate to professional secrecy and quality of legal advice as follows.
Only natural and legal persons can be held liable under Swiss private law. When using AI-powered products or services, the key question that arises is therefore which person the liability for the conduct of the AI is attributed to. Liability for damage or losses caused can be either contractual or non-contractual (tortious).
Contractual liability
Pursuant to Article 97 of the Swiss Code of Obligations (CO), a party to a contract is liable for failing to fulfil its contractual obligations unless it can prove that it was not at fault (ie, that the failure occurred without intent or negligence). Pursuant to Article 101 CO, the contractual liability extends to the conduct of any auxiliaries, such as employees or subcontractors, that are used in the performance of the contractual obligations; only natural or legal persons can qualify as auxiliaries. If the use of AI leads to a breach of a contractual obligation, the key question will be whether the breaching party can prove it did not act negligently in its use of AI.
Non-contractual (tort) liability
Article 41 CO covers non-contractual liability, where any person who unlawfully causes damage to another person is obliged to provide compensation. The application of this provision further requires intent or negligence on the part of the person causing the damage. Article 55 CO extends this liability to an employer for the conduct of their employees or other auxiliary staff.
This liability regime is supplemented by the Product Liability Act, which imposes liability on the manufacturer of defective products. However, there is no established conclusion on whether software such as AI systems qualify as products within the meaning of the Product Liability Act, and to what extent an individual AI system or component could be considered defective while the technology as such is known to produce unintended results. Furthermore, the manufacturer of a defective product is not liable if they can prove that the defect could not have been detected based on the state of the art in science and technology at the time the product was placed on the market.
There is no applicable information in this jurisdiction.
At present, there is no specific regulation on algorithmic bias in Switzerland. However, the fundamental right to equal treatment and non-discrimination contained in the Federal Constitution applies throughout the Swiss legal system and can also affect the relationship between private parties (eg, between employee and employer) in addition to the state.
Please refer to 8.3 Data Protection and Generative AI for general remarks on the lawfulness of processing and to 11.4 Automated Decision-Making for the consequences of fully automated decisions.
One key point in which the Federal Act on Data Protection (FADP) differs from other data protection laws is that fines for criminal offences of up to CHF250,000 are imposed on the responsible individuals (natural persons), not on the company. The main offences under the FADP include the intentional violation of information and cooperation obligations, of duties of care (eg, in connection with delegation to a processor), and of data protection secrecy obligations.
While it is unlikely that the use of AI as such will constitute a criminal offence under the FADP, the individuals responsible for the design and implementation of data processing using AI should carefully review data protection compliance to avoid personal criminal responsibility.
Biometric information that uniquely identifies a natural person expressly qualifies as “sensitive personal data” under the Federal Act on Data Protection, which means that such data may not be disclosed to third-party controllers without justification and, where the consent of the data subject is required for processing, the consent must be explicit.
The public’s unease regarding facial recognition and biometrics allegedly already in use by the police forces in certain Cantons has led to legislative initiatives in multiple Cantons to expressly prohibit their use.
Automated decision-making is addressed by the Federal Act on Data Protection. Its Article 21 obliges a data controller to inform the data subject if a decision is made solely by the automated processing of personal data that has a legal or otherwise significant consequence for the data subject. This obligation does not apply in the context of entering into an agreement if the request of the individual is granted in full (eg, a loan application is fully approved by an automated decision). Where the data subject has to be informed, they are entitled to request a human to review the automated decision. Intentional omission of the information of the data subject under Article 21 constitutes a criminal offence for which a fine of up to CHF250,000 may be imposed on the responsible individual.
Article 21 is not applicable to decisions recommended by AI but manually approved by a human, even if the level of human review is superficial. There is a pending initiative in the Swiss parliament to include similar information obligations for any decision significantly based on AI, but it is likely that further consideration will be postponed until the Federal Council’s review of potential approaches to general AI regulation is completed (see 3.1 General Approach to AI-Specific Legislation).
There is no general obligation for businesses to disclose the use of AI in Switzerland. An obligation of transparency may exist with respect to individual uses of AI, such as:
Many antitrust implications of price-setting using AI are complex and not yet clearly established under Swiss antitrust law. While it is clear that the use of AI to implement pre-agreed horizontal price fixing remains illegal, the possibility of AI autonomously coordinating prices among competitors is being discussed among scholars, but there is no settled case law yet.
However, in a vertical price-fixing case, the Swiss Federal Supreme Court held that the communication of manufacturer-suggested retail prices by the manufacturer to its distributors through an automated electronic system with daily price updates constituted an illegal agreement on fixed prices, seeing as all distributors and retailers had reason to believe that these prices would also be used by their competitors (and they in fact complied with the suggested pricing).
As AI models are only as good as the data they were trained with, businesses procuring AI solutions intended for productive use should include provisions regarding the quality of the training data and corresponding properties of the AI model (eg, no discriminatory bias, no infringement of IP rights) in their procurement contracts. Otherwise, they themselves risk becoming liable for the output of the AI solutions. While the use of due care is a possible defence under most provisions giving rise to liability under Swiss law, businesses ignoring well-known risks inherent in current AI technology may face the accusation of not having used the required standard of care.
Hiring
Résumé screening software (with or without AI) has been used in Switzerland for some time. Under the principle of freedom of contract, private-sector employers in Switzerland are not required to hire individual applicants, and, by default, are not required to explain why an applicant was rejected.
However, under the Gender Equality Act, employers are not permitted to discriminate against applicants based on gender in the selection process: if an applicant is rejected due gender-based discrimination, the applicant is entitled to a written explanation for the rejection and, if they can prove discrimination, to compensation under Article 5(2) of the Gender Equality Act. However, this does not apply at the stage of job ads, where employers are permitted to advertise a position only for a specific gender. Where AI-based selection tools discriminate based on gender, for example due to inherent gender bias in the training data, the employer may become liable.
In addition, if any rejection of an applicant is made by an automated decision without human control, the applicant must be informed, and is entitled to have a human review the decision based on data protection law (see 11.4 Automated Decision-Making).
Termination
Termination of a private-sector employment relationship is governed primarily by the Swiss Code of Obligations. In principle, employers do not need specific legal grounds for termination, but do need to provide a written explanation upon request of the employee. However, Article 336 of the Swiss Code of Obligations prohibits wrongful termination, including termination due to a quality inherent in the employee’s profile (eg, age, ethnicity, religion), unless this quality is related to the employment relationship or significantly affects work in the business. If an employee is terminated based on recommendations by AI (eg, due to inherent bias in the training data), the employer may therefore become liable for compensation due to wrongful termination (up to six months’ salary).
The prohibition of gender-based discrimination under the Gender Equality Act and the provisions on automated decision-making also apply to any termination that is discriminatory or automated, respectively.
Monitoring and surveillance is considered detrimental to employee health, and is therefore subject to multiple restrictions. Article 328 of the Swiss Code of Obligations provides for a general duty of care of the employer to respect and protect the employee’s personality rights and health. Based on this principle, all measures taken that can affect employee health and wellbeing must be proportionate to the (legitimate) aim pursued. In addition, Article 26 of Ordinance 3 to the Employment Act specifically prohibits the use of systems to monitor the general behaviour of employees in the workplace. Such systems are only permitted where they pursue legitimate aims, such as security, performance or efficiency, and only if their use is proportionate. In practice, this means employers must be able to prove that less intrusive alternative solutions were not sufficient to achieve the aim. Specifically, the FDPIC has stated that AI-based systems for the automated evaluation of employee-based data (eg, vision, movement, communication patterns) are prohibited under these provisions.
Employees are also entitled to employee participation regarding all matters that concern the protection of their health. They must be sufficiently informed of measures and circumstances affecting their health, and have a right to be consulted before the employer takes relevant decisions. This includes the right of employees to make suggestions, and the obligation of the employer to explain decisions that do not take employees’ objections into account.
At present, there is no specific regulation regarding the use of AI in digital platform companies, and general observations regarding the use of AI in data protection and potentially in employment contexts also apply to digital platform companies. Extensive use of AI to control the conduct of platform participants (eg, which participants get to serve which customers when, and at what price) may lead to increased findings by courts that the platform participants lack autonomy and are in fact employees of the digital platform companies.
At present, there is no specific regulation regarding the use of AI in financial services. General financial services regulation, including with respect to risk management, also applies to the use of AI. As set out under 5.1 Regulatory Agencies, FINMA has identified AI as a trend in its risk monitoring report, and will review the use of AI by supervised institutions.
At present, there is no specific regulation regarding the use of AI in healthcare. The primary areas of existing law governing the use of AI in healthcare are sector-specific regulation, such as the Therapeutic Products Act and the Medical Devices Ordinance, which specifically includes software in its definition of medical devices. Where AI-based products qualify as medical devices, they need to comply with the general safety and performance requirements set out in Annex I to the EU’s Medical Device Regulation (by virtue of an express reference to EU legislation). Medical devices placed on the market in Switzerland require a conformity marking (CE label), predominantly based on self-assessment by the manufacturer. The relevant assessment criteria depend on the applicable risk category.
In addition, the Federal Act on Data Protection applies also in the healthcare sector, although in the area of research on humans it is supplemented and partially overridden by the Human Research Act.
Fully autonomous vehicles are not yet permitted on Switzerland’s streets (outside of specific, limited pilot experiments), and drivers must always have their hands on the steering wheel. However, the Swiss parliament amended the Road Traffic Act in 2023 to permit the Federal Council to authorise vehicles with automatic driving systems in a delegated act. In the near future, drivers of such vehicles may therefore let go of the steering wheel, provided they remain ready to resume operating the vehicle themselves if the system indicates a need for manual intervention or otherwise reaches its limits.
At present, there are no specific regulations regarding the use of AI in manufacturing. The liability aspects of manufacturing primarily depend on whether there is a contractual relationship or not (see 10.1 Theories of Liability). Where the use of AI causes manufactured products to be defective, this may trigger liability of the manufacturer under the Product Liability Act.
Please refer to 9.1 AI in the Legal Profession and Ethical Considerations for the main concerns. Many professional services occupations are subject to similar professional secrecy obligations as lawyers (accountants, doctors, other medical professionals, etc.), and most professional services firms are concerned about the confidentiality of the information provided by their clients as well as the accuracy of their advice.
At present, there are no reported judicial or agency decisions in Switzerland on whether AI technology can be an inventor for patent purposes or an author for copyright purposes. However, the prevailing doctrine among legal scholars is that AI can neither be an inventor nor an author.
Patent law
Article 3 of the Swiss Patent Act refers to an “inventor” being entitled to a patent. Prevailing doctrine in Switzerland provides that only a natural person (ie, a human) can be an inventor for the purposes of the Patent Act, which excludes both legal persons and AI. Switzerland is party to the European Patent Convention (EPC), which contains analogous wording in its Article 60. With respect to the EPC, the European Patent Office’s Legal Board of Appeal ruled on 21 December 2021 in the DABUS case (case no. J 0008/20) that AI may not be named as an inventor.
The foregoing applies to the extent AI itself generates the invention without significant human contribution. If a human uses an AI application solely as a tool to make a discovery, the same way they might use other computer software, the human using AI as a tool would be considered the inventor under the Patent Act, and the invention may be eligible for patent protection.
Copyright law
Article 2 of the Swiss Copyright Act defines a work in which copyright may subsist as an “intellectual creation of literature or the arts with individual character”. The criterion of “intellectual creation” is widely interpreted as an expression of the human mind. Consistent with this interpretation, Article 6 of the Copyright Act states that “the author is the natural person who has created the work”. A work within the meaning of copyright can therefore only be created by a human, and not by AI.
While AI may not be an author and purely AI-generated content can therefore not be protected by copyright in Switzerland, works created using AI may be protected. If AI is used solely as a tool by the human author to express their thoughts, for example by having generative AI translate or edit existing text of the author or by modifying pictures similar to what an artist might do with photo editing software, the resulting AI-generated work may still be subject to copyright protection. Equally, if a human sufficiently modifies purely AI-generated content or is creative in the selection or ordering of pieces of AI-generated content, the resulting overall work may also be protected by copyright.
Swiss law uses the terms “manufacturing secrets” and “business secrets”, rather than trade secrets, but there is no distinction in practice. While there is no statutory definition, case law defines trade secrets as any information that: (1) is not publicly known, (2) has commercial value, (3) the owner has a legitimate interest in keeping secret, and (4) the owner intends to keep secret. A curated training data set, an AI model, and the source code of an AI application can therefore constitute a trade secret if the foregoing criteria are met.
Trade secrets are not protected as an absolute right (such as copyright) under Swiss law, but their disclosure or misappropriation is prohibited under specific circumstances, as follows.
All examples of unfair competition above can trigger both civil and criminal liability.
See 15.1 Applicability of Patent and Copyright Law.
Using OpenAI’s products, particularly ChatGPT, touches on multiple issues relating to intellectual property, both on the “input” and on the “output” side, as follows.
Pursuant to Article 716 of the Code of Obligations, the board of directors of a Swiss corporation has the non-transferable and inalienable duty to set the company’s overall strategy, to determine the company’s organisation and to supervise the executive officers.
When advising corporate boards of directors on identifying and mitigating risks associated with the adoption of AI, several key issues should therefore be addressed. These include the following.
Given that Switzerland does not have AI-specific regulation at present, compliance strategies should be aimed at ensuring that the use of AI tools complies with the existing body of law, and should therefore focus on practical guidance for employees to use AI tools. Businesses should ideally adopt internal policies on the use of AI tools, clearly outlining which types of use are permitted and which are not.
Depending on the field of business and the contractual frameworks in place, this may include an outright prohibition on using publicly available generative AI tools (eg, public versions of ChatGPT), restrictions on the type of data that may be submitted as input (eg, no confidential information, no personal data) or restrictions on the use of output (eg, mandatory review for accuracy prior to publication).
Businesses should also invest in training their staff in AI literacy to help avoid issues based on ignorance or misunderstanding the nature and limitations of current AI technologies (eg, mistaking a generative AI chatbot for a search engine).
Prime Tower
Hardstrasse 201
CH-8005 Zurich
Switzerland
+41 43 222 10 00
+41 43 222 15 00
lawyers@homburger.ch www.homburger.ch