Contributed By Sołtysiński Kawecki & Szlęzak
Regulation (EU) 2024/1689 of the European Parliament and of the Council of 13 June 2024 laying down harmonised rules on artificial intelligence and amending Regulations (EC) No 300/2008, (EU) No 167/2013, (EU) No 168/2013, (EU) 2018/858, (EU) 2018/1139 and (EU) 2019/2144 and Directives 2014/90/EU, (EU) 2016/797 and (EU) 2020/1828 (Artificial Intelligence Act; AI Act) applies directly in Poland. There are no general provisions of Polish law that would specifically apply to AI. However, the draft of the Act on Artificial Intelligence Systems is being prepared by the government in order to ensure full implementation of the AI Act.
AI is currently qualified as software, and in addition to the AI Act, the laws applicable to software will apply to AI. The following are examples thereof.
AI and machine learning (ML) have been applied in various sectors in Poland, but AI deployments are rather slow and focus on automation and efficiency.
From the authors’ experience, key industry applications are now based on generative AI (GAI) and include the following.
There are multiple cross-industry initiatives concerning new technologies in Poland.
The Polish government’s initiatives to facilitate and support the adoption and advancement of AI for industry use are limited and include the following.
See also 7.2 Judicial Decisions.
There are no AI-specific regulations in Poland. However, work is in progress at the government level on an act implementing the AI Act – see 3.7 Proposed AI-Specific Legislation and Regulations.
No AI-specific legislation has been enacted in Poland yet.
AI Policy
AI Policy focuses on creating transparent and accountable algorithms for use in public administration, enhancing data access, and applying AI to healthcare and environmental protection. In February 2024, the works on update of the AI Policy started.
Position of the Polish Financial Supervision Authority (FSA) on the provision of robo-advisory services (2020)
The guidelines emphasise the user’s control over AI use and responsibility for clear client communication. Humans should make the final decision.
Recommendations on AI in the financial sector (2022)
The Ministry’s Working Group on AI – Subgroup for the Financial Sector – identified several barriers to using AI and provided its recommendations in the identified fields.
Recommendations for the use of AI in justice and law enforcement (2024)
The document suggests adopting AI to modernise and speed up the judicial system – including digitalising records, automating transcriptions, drafting orders, using chatbots, searching for case precedents, drafting decisions, and implementing electronic delivery and translation systems.
Guidelines on the responsible use of generative AI in research (2024)
In March 2024, the European Commission – together with the European Research Area countries and stakeholders – published guidelines focusing on research quality, honest GAI use, respect for participants, and accountability in research.
Recommendations on prohibited AI systems
In September 2024 Ministry of Digitalisation and the Research and Academic Computer Network – National Research Institute published their recommendations on prohibited AI systems and associated provisions of AI Act.
The publication aims to clarify the provisions of the Act regarding prohibited artificial intelligence systems, detailing the techniques, functions, and methods by which these systems operate.
Policy for the Development of Artificial Intelligence in Poland 2025–2030
Prepared by the Working Group on AI in December 2024 to develop trustworthy AI in Poland, focusing on four pillars: human capital, innovation, investment and implementation. The study also takes into account important areas of development, such as economic competitiveness and national security, and invites discussion on the future of AI in Poland.
Recommendations on how attorneys-at-law should use AI-based tools
Launched by the National Council of Attorneys-at-law in May 2025, the Recommendations are the first extensive document with guidelines for the Polish legal community on the use of AI. The document includes recommendations for three phases: preparation, implementation and use of AI tools.
Similar to other EU member states, the regulatory framework for AI in Poland is predominantly defined by the AI Act. This framework is further supported by the EU Directive on liability for defective products (see 10.2 Regulatory), in addition to sector-specific EU product regulations.
Poland currently has no AI-specific laws, resulting in an absence of potential inconsistencies with EU law. The proposed Polish Act on AI Systems is to ensure enforcement of the AI Act in Poland (see 3.7 Proposed AI-Specific Legislation and Regulations).
This is only applicable to the US.
No special local Polish laws have been introduced or amended to foster AI development. Implementation of Directive on copyright and related rights in the Digital Single Market, introducing text and data mining exception, was delayed and came into force only in September 2024. Two separate exceptions are established: (i) for cultural heritage institutions and certain educational entities, and (ii) for any individual or entity, although rights-holders may opt-out in such case. As of now, no recognised national standard or recommendations concerning the opt-out mechanism have been developed.
While the EU’s primary aim is to regulate AI through the EU AI Act, AI must also comply with all other EU regulations, ie:
Works on the draft Polish Act on AI Systems are ongoing at the government level. The published draft of the Act includes provisions on the following.
See also 5. AI Regulatory Oversight.
There have been no decisions related to AI in Poland. However, Poland is introducing new digital tools to support judges and improve the efficiency of the court system. The main initiative is the Digital Judicial Assistant (DJA), which will help judges write the reasoning behind court decisions. DJA will be available in selected courts starting mid-2026, and judges will receive training on how to use it. DJA is initially foreseen to be used in the Swiss franc loan cases. The project also includes the so-called Settlement Calculator, ie, the tool that will automatically calculate what each side (the consumer and the bank) owes in mortgage cases, especially those involving Swiss franc loans. These innovations are part of a broader effort to modernise the Polish judiciary and make it more efficient, accurate, and user-friendly. Currently, DJA is tested in the District Court in Poznań but its function is limited to assist administrative secretaries.
EU
Poland
Additionally, several governmental bodies and institutions play roles in regulating aspects related to AI. Key entities are:
A legal definition of AI was adopted in the AI Act.
The European Commission approved the content of draft:
These guidelines are not legally binding and may be updated in the future in response to new practices and insights.
So far, the enforcement and other regulatory actions have limited scope. For instance, UODO is investigating a complaint about ChatGPT (see 8.2 Data Protection and Generative AI).
The Polish Committee for Standardisation (PKN), which creates and approves Polish Standards (PN), established a separate Technical Committee on AI. It is the lead Polish committee for international co-operation in international standardisation committees: ISO/IEC JTC 1/SC 42 AI and CEN/CLC/JTC 21 AI.
Currently, the committee’s work agenda mainly includes translations of existing international standards. In view of the standardisation works at the European level (in relation to AI Act requirements), the significance of national standards will most probably be limited.
See also information about the AI Policy – 2.2 Involvement of Governments in AI Innovation.
The standards are usually intended for voluntary use by businesses, but in some sectors, compliance with such standards is deemed essential for the credibility and reliability of products.
European
Due to abstract nature of requirements imposed on providers of high-risk AI systems under the AI Act, harmonised technical standardisation will play a key role in operationalisation of these requirements.
European Commission issued a standardisation request to the European Committee for Standardization (CEN) and the European Committee for Electrotechnical Standardization (CENELEC). A joint committee by CEN and CENELEC (CEN-CENELEC JTC 21) is planning to provide more than 30 technical standards in relation to the AI Act.
In view of the delayed standardisation works at the European level and the upcoming entry into force of obligations of providers of certain high-risk systems under the AI Act (August 2026), implementation of the expected technical standards may prove challenging, especially for startups and SMEs.
International
Polish government is engaging with AI across various domains, and the known examples include:
Digital Judicial Assistant (See 4.1 Judicial Decisions)
“Bielik” LLM was created as a result of the work of a team operating within the SpeakLeash Foundation and the Academic Computer Centre Cyfronet of AGH Academy of Science and Technology as Polish large language model. AI is not widely used in the criminal justice system, but its use is thought to be very widespread. Currently, the most common example of using AI in this area is a general legal information system at the disposal of judges and prosecutors. This system includes a search system (for rulings, literature, etc) based on AI algorithms.
This is not applicable in Poland.
In March 2025 the Ministry of Justice announced launching the “Digital Court”, the project that will encompass introduction of:
The AI Policy (see 2.2 Involvement of Governments in AI Innovation) emphasises the importance of AI in national security and encourages co-operation between the private and military sectors to address defence needs.
According to the recommendation to the National Security Strategy of the Republic of Poland (2024), AI creates new development opportunities for Poland while generating previously unknown risks, including loss of digital sovereignty. The recommendation highlights the importance of monitoring and responding to the development of emerging technologies, such as artificial intelligence and quantum technologies, within the context of security and international relations. Additionally, it emphasises promoting STEM (Science, Technology, Engineering, and Mathematics) education to improve technological competencies.
The GAI raises many issues, the key (at the moment) including:
Copyright Protection
AI-generated results
Copyright protection of AI-generated output is limited since only humans can be considered authors under Polish law (as in most countries). The AI output cannot be considered as copyrightable work.
It may be claimed that AI is only a tool and AI-generated output can be copyrightable work as long as human oversight in its creation was significant and that all creative choices and decisions were made by a human, as with photography. This can apply only to specific examples and should be assessed case by case.
It’s worth noting that this perspective is gaining more support among legal practitioners compared to the early stages of this discussion.
With regard to derivative (neighbouring) rights, such as rights to recordings of video or sound (as protection may be granted, eg, to birds’ sounds), the rights may be allocated to individuals or companies.
The position with respect to the output ownership depends on the market. For example, MidJourney T&Cs provide that users of the commercial (paid) version of the platform are the sole owners of the content, while free version users only receive the licence for the personal use of such content. Users are responsible for the input and are obliged to indemnify MidJourney if there is any third-party rights violation due to such non-compliant input. Microsoft® Product Terms state that Microsoft® does not own the output of Generative AI Services.
The output may infringe on copyrights, other IP rights, or third parties. Therefore, some providers are following Microsoft’s® customer copyright commitment to offer indemnification should a third party raise a claim that AI output infringes their IP rights. Such indemnification is usually subject to certain conditions (eg, using filters or content monitoring tools).
Moreover, the providers usually specify in their T&C if they use the output for training the model.
AI input
The input provided to AI systems can be protected as a standard work under copyright law, as long as it meets all the copyright law requirements (eg, originality and creativity). T&Cs usually specify whether the input will be used to train model.
AI system itself
An AI system itself can be protected under copyright law as the software. In Poland, as in the whole EU, software is protected similarly to literature works, with some necessary modifications due to the nature of the computer programs.
Training data
Training data can be protected by copyright or by databases protection laws. However, there are discussions about whether the use of training data will be an infringement of copyright or may be based on text and data mining exemptions (once implemented in Poland). There are arguments raised that training the AI models is similar to learning from a book and consists of learning ideas or concepts that are not protected by copyright. However, no court cases have been raised in Poland against the model developers.
Trade secrets; know-how
Input, training data, software, and output can be considered trade secrets as long as the definition of the trade secret is met (eg, they have economic value, are kept confidential, and security measures are introduced to maintain their confidentiality). This may be relevant in cases where AI output cannot be protected (eg, software output).
The main data protection issues connected to GAI are:
If possible, the rights of data subjects should be respected. However, the difficulty in identifying and accessing personal data used to train the AI or in some cases – stored in AI models can be difficult. The risk of not being able to exercise the rights of the data subject can be minimised by implementing measures such as keeping a special register of traceable data, maintaining a management system that allows tracking the use of data and implementing data minimisation techniques at an early stage of system design. In particular, a request to delete data should not entail the deletion of the whole model. However, the model should be reviewed in order not to create future output with false data or data of a person who requested to have their data deleted. The right to rectification applies to both system input and output data. Consequently, the possibility of exercising this right must be taken into account at the AI design stage to avoid the risk of retraining the AI based on non-rectified data.
Guidance in this scope is expected following a complaint filed in Poland in September 2023 claiming a lack of data subject rights fulfilment and transparency in ChatGPT.
The National Bar Council of Attorneys-at-Law (Krajowa Izba Radców Prawnych) has prepared recommendations for attorneys-at-law on the safe use of AI in legal work, including recommendations on how to prepare for the use of AI tools, on the implementation and application of AI tools. As planned, the recommendations will be published in mid-May 2025.
There are no case law so far.
As for AI solutions, most law firms use standard AI tools, particularly for translation purposes or for achieving efficiency and improving the quality of work (eg, summarisation of meetings, drafting clauses, etc). The most common seems to be Copilot (including Microsoft 365® Copilot). AI tools are also used in e-discovery to identify relevant documents. Other solutions (eg, supporting drafting litigation documents) are often not adjusted to Polish law and, therefore, are of limited use.
The legal practitioners should ensure that:
A legal adviser is obligated to observe professional secrecy and ensure that appropriate technical and organisational measures are in place to safeguard against its disclosure. Legal advisers should therefore ensure that AI tools are safe, among other things, by selecting a reliable AI tool provider, as well as assessing whether the use of a particular AI tool requires additional arrangements with client.
Currently, the use of AI is subject to standard liability rules set forth in national laws.
Fault-Based Liability
The aggrieved persons need to prove the fault of the liable person, the damage and the causation between the fault and the damage to successfully bring a liability claim. In the context of AI, reliance on this principle is difficult, as often it may not be easy to find the person responsible or the cause (taking into account the lack of access to the model’s “construction”). Compliance with AI manufacturer’s instructions may shield users from fault while altering an AI code or using it for unintended purposes could attribute fault to the user.
Strict (Risk-Based) Liability
In specific cases, Polish law attributes liability to certain persons without the need of the aggrieved person to prove fault. An individual who operates an enterprise powered by natural forces (such as steam, gas, electricity, or liquid fuels) is liable for injuries or property damage resulting from the operation without the necessity to prove the fault. However, this provision does not apply in the case of AI, as providers of AI systems or models will not qualify as operators of enterprises powered by natural forces.
The risk liability also applies to the vehicle’s possessor, though it is still necessary to prove the cause. Thus, its application will still be difficult if the damage caused by the AI system is applied to the vehicle.
Liability for Defective Products
The current rules, based on the implementation of Council Directive 85/374/EEC, do not allow considering AI systems themselves as “products” within the scope of provisions related to defective products as they are exclusively movable items. AI systems cannot be classified as a movable product. However, please see section 10.2 Regulatory.
Contractual Liability
If the aggrieved party is using AI based on the contract, it may potentially claim the damage for breach of contract; it has to prove the breach, the damage and causation. There is a presumption that the alleged perpetrator is liable for the damage. However, they may prove otherwise. The risks outlined above to claim the damage caused by AI also apply to the contractual liability. However, the provider of AI solutions often offers additional measures, such as indemnity for third-party claims related to the infringement of IP rights by output generated by provided AI. On the other hand, in most business transactions, the parties limit the provider’s liability (eg, up to 12 months’ remuneration). The limitations or exclusion of liability will not be effective in contracts with consumers or the case of B2B if the damage is caused wilfully.
Liability for Infringement of Personal Interests
This is a fault-based liability. It is attributable to the person who uses the AI systems or the creations of AI systems without proper verification or with a false intent, eg, using deepfake technology. Besides the typical financial claims, the plaintiff may request the infringer to publish a given statement in the media.
Insurance Position
At the moment, there is no obligatory insurance similar to that necessary for the use of vehicles. Due to the confidentiality of model training, it is rather unlikely that insurance coverage will be an affordable standard solution in the foreseeable future.
Directive 2024/2853 on the Liability for Defective Products Repealing Council Directive 85/374/EEC
The directive entered into force on 8 December 2024 and member states should make their legal systems compliant by 9 December 2026. While similar to the existing Directive on defective products from 1985, it introduces changes that could significantly impact the liability of AI systems.
It will cover not only physical goods but also software (especially damages for destroyed or corrupted data). The burden of proof will be simplified, and the aggrieved party will be able to claim material and mental damages (confirmed medically).
The Directive of the European Parliament and of the Council on Adapting Non-Contractual Civil Liability Rules to AI
The European Commission decided to withdraw the proposal of Directive in February 2025, but some members of European Parliament criticised it and discussions continue. Thus, the proposal has not yet been formally abandoned.
Bias Characterisations and Risks
Algorithmic bias happens when the system discriminates against a specific group or individual, resulting from various factors, eg, biased training data, discrimination in data collection, a biased training team or defective parameters in the models, or inappropriate deployment. Bias has not been explicitly defined in the Polish legal system. AI bias can affect the personal interests and freedoms of individuals, for example, by discriminating against them in a recruitment process or credit scoring, which may lead to claims for compensation and erosion of consumer trust.
Regulations and Industry Efforts
The GDPR introduces a human oversight requirement for processes that qualify as automated decision-making unless required by law. Human oversight can take different forms:
See also 13. AI in Employment.
Under Article 10 of the AI Act, in the development of high-risk AI systems that use data for training, validation and testing, it is crucial to adhere to quality criteria, such as data governance, examination for biases that are likely to affect the health and safety of persons or have a negative impact on fundamental rights or lead to discrimination prohibited under EU law.
Training, validation and testing data sets shall be relevant, sufficiently representative, and to the best extent possible, free of errors and complete in view of the intended purpose. The data sets shall have the appropriate statistical properties, including, where applicable, as regards the persons or groups of persons in relation to whom the high-risk AI system is intended to be used. Even when an AI system isn’t deemed high-risk, it remains crucial for organisations to carry out their own risk evaluations to mitigate any potential adverse outcomes, including bias.
Transparency requirements arising from Article 13 of the AI Act also aim to ensure clarity and minimise bias in AI.
Article 14 of the AI Act requires that high-risk AI systems that continue to learn after being placed on the market or put into service shall be developed in such a way as to eliminate or reduce as far as possible the risk of possibly biased outputs influencing input for future operations (feedback loops), and to ensure that any such feedback loops are duly addressed with appropriate mitigation measures.
Legal Issues – Liability
The AI Act provides restrictions on using real-time and retrospective facial recognition, effective from February 2025:
The use of such systems is also subject to GDPR and local laws implementing the Data Protection Law Enforcement Directive (LED).
Risks
Automated Decision-Making (ADM) employs algorithms and AI systems to automate tasks traditionally requiring human insight, learning from data to predict outcomes.
Risks
In terms of risks, ADM systems can inherit biases (section 11.1 Algorithmic Bias). ADM models often lack transparency, leading to mistrust.
Applicable Rules – Article 22 of the GDPR
European Parliament Resolution on Automated Decision-Making Processes indicates that when consumers are interacting with ADM, they should be properly informed how it functions, how affects a human with decision-making powers, and about how the system’s decisions can be checked and corrected.
DSA
AI Act
The principle of transparency is one of the basic principles introduced by the AI Act. This principle will be applied from 2 August 2026.
Providers of general-purpose AI models are required to make publicly available a detailed summary of the content used to train the model, particularly text and copyrighted data.
The procurement of AI technology in Poland can be implemented by two groups of entities: private entities or public entities.
Private Entities
In cases of private entities, the principle of freedom of contract applies. The current key issues which would have to be addressed in contracts for procurement of AI solutions include the following: whether the input or output data are used to train the models, ownership of output, liability and warranties, IP-related issues (in particular, dealing with third-party claims to the model or output), use of personal data (processor, controller, joint controller), data flows, location of input, output and model, abuse monitoring and content filtering, possibility of suspension of services in cases of abusing services or due to political tensions, supply chains.
Public Entities
Public entities may be obliged to conclude the contract for AI technology under public procurement rules, and they often provide a template contract in the contracting documents. As they are not experts in AI, the proposed contracts may not address relevant risks and may not be consistent with market standards. Therefore, it is recommended that public entities carry out two-step procedures or negotiations to first precisely identify its goals and the way in which they may be achieved.
Employers are not allowed to use AI systems to reach conclusions about a job candidate’s/employee’s emotions, except where it is intended to be put in place or into the market for medical or safety reasons .
The employer will have additional obligations when using following AI systems, which are classified as high-risk AI:
Employers utilise tech and software for recruitment and termination processes to save costs and time. Notwithstanding the regulation of the AI Act, employers must adhere to general anti-discrimination and privacy laws:
Employers use various digital tools to evaluate employees’ performance, track working time, and review employees’ work or use of resources. There are also AI tools that can provide real-time feedback on an ongoing basis, informing employees of their progress and areas for improvement. If the technology is used to monitor employees’ email and/or activity, such monitoring should be introduced in a formal way via relevant internal policies and procedures, announced to employees, and applied within their frameworks. It cannot infringe on the integrity of correspondence or on an employee’s personal interests.
The breach of the rules may result in employees’ claims for breach of their rights by illegal monitoring and compensation. It may also constitute a breach of privacy laws.
Using AI on digital platforms should be assessed in light of the Digital Services Act (DSA). Most of all, the following conclusions shall be kept in mind:
There are no specific rules regarding employment in Digital Platform Companies (see also section 13. AI in Employment).
Financial services institutions (FSI) are very focused on innovation, including AI. For example, a Polish bank implemented a chat service for customers, which informs customers about the bank’s services and analyses customers’ transactional data. FSI are also interested in services that automate internal activities, eg, anti-money laundering and anti-fraud purposes.
There are no legal provisions regarding specifically AI in FSI (except for application of AI in credit scoring systems, which may fall into the AI Act high-risk systems rules). However, financial services are highly regulated and introducing new technologies is usually subject to complex obligations:
New technologies may be highly beneficial for the healthcare sector. ML algorithms may analyse big datasets and entail accuracy and speed in healthcare providers/health research activity. ML and AI can assist in diagnosis and disease management, interpreting medical images, drug discovery, operations. Also, new technology may enable the use of new ways of collecting data and providing health assistance, eg, telemedicine/chats or wearable technology. However, especially in Polish public sector, the adoption of AI in healthcare is limited by the restrictions in using the cloud computing through which such AI solutions are often offered. Some restrictions in this area were removed in October 2024 but the works still continue.
Regulations and Main Issues
There are no regulations related specifically to using AI in healthcare and connected ML. The general regulations will apply, ie:
Risks
Polish law generally prohibits the use of autonomous vehicles (AVs); however, automation features under the supervision of a human driver can be used. There is an exemption for approved research testing on public roads, but not every type (level) of an AV can currently be tested on Polish roads – only vehicles at level 3 of automation can be tested in Poland. There are no AI-specific regulations, however the Road Traffic Act provides a definition for the purposes of conducting controlled AVs testing, under which AV means a motor vehicle fitted with systems which control the movement of the vehicle and enable it to move without intervention from the driver, who may take control of the vehicle at any time. The Ministry of Infrastructure intends to facilitate testing autonomous vehicles on all levels of automation and their equipment on public roads in near future.
Liability
The standard vehicle liability rules also apply to AVs (see section 10. Liability for AI). Motor vehicle users are required to hold mandatory insurance for civil liability.
Significant Issues
In the AI Policy, it is acknowledged that AI-based systems used in autonomous transport can significantly lower the number of accidents – and, as a result – the number of fatalities. However, using AVs is linked to certain issues and risks, ie:
There are no specific regulations related to using AI in manufacturing – the rules related to liability are described in section 10. Liability for AI and the employment issues are described in section 13. AI in Employment. Data privacy applies. The rules regarding product safety are based on the Machinery Regulation, which generally aims to cover new technologies. AI may be beneficial to reduce the size of the workforce. However, as AI will automate certain activities, it may lead to layoffs, increased unemployment and loss of competencies. Manufacturers may be one of the key sources of data for training AI based on the Data Act.
There are no specific regulations related to using AI in professional services; thus, the rules related to liability described in section 10. Liability for AI will apply. Providing some professional services is covered by sector-specific regulations and ethical rules (eg, legal services) and using AI should enable compliance with them. Most importantly:
There are no binding laws or regulatory intentions with regards to IP and AI in Poland. There are also no administrative or judicial decisions in this respect. See 8.1 Specific Issues in Generative AI for general remarks regarding the relation between IP and generative AI.
There are no decisions in Poland relating to whether AI technology can be an inventor or co-inventor for patent purposes or an author or co-author for copyright purposes.
The prevailing view is similar to one presented already in the USA or other countries in the EU, ie, that AI-created works or inventions cannot be copyrighted or protected by industrial property law.
Patent Law
With regard to patent protection, only a human can be an inventor. It was confirmed in the DABUS case in which both the European Patent Office and later the Board of Appeal refused patent protection for an AI system.
Copyright Law
In Poland, copyright protection arises by operation of law. There is no need to file an application, and there is no Copyright Office. Disputes relating to copyright infringements are decided by special IP departments of common courts.
Recently, in the EU, the first case regarding copyright ownership of AI creations was decided when a Prague court ruled that DALL-E creations aren’t copyrightable as they’re not human-made. Additionally, the Union of Authors and Composers for the Performing Arts in Poland, the largest copyright collective management organisation, has amended its regulation on submitting and registering works. Under this regulation, AI-created works are generally excluded, but it is possible to register a work co-created with AI. In such cases, the human creative input must be quantified by a percentage indicator.
Contractual Provisions
Non-disclosure arrangements are the common legal instruments to protect AI technologies and generated content. However, it is limited to the parties and is not generally enforceable like IP protection, which applies to all potential violators.
Trade Secrets
AI technology, training data, input, or output may be protected as trade secrets, as they do not necessitate human authorship. As confidentiality is one condition of such protection, appropriate instruments must be in place (eg, non-disclosure contracts).
The prevailing view is that the creator of copyright-protected works can be human only; however, video or sound output may be protected as derivative rights. Additionally, AI systems can be regarded as tools in the creation process, provided that the human contribution is significant enough.
See 8.1 Specific Issues in Generative AI.
Using OpenAI while creating works and products is generally associated with the same risks as using other GAI systems.
The Polish authorities do not currently use AI in assessment of their cases. They have not seen AI-related transactions or practice either.
Currently in Poland, there is the National Cyber Security System Act (NCSSA), which defines the organisation of the national cybersecurity system and the tasks and responsibilities of the entities that are part of this system. The legislation is undergoing legislative work aimed at increasing the level of cybersecurity in Poland – including addressing challenges related to AI. The new regulations will be flexible and applicable to increasing the scope and speed of malicious actions, including those involving AI and LLMs usage. The amendments to the NCSSA are going to implement the EU NIS-2 Directive, which provides new cybersecurity requirements and extends obligations to many entities from various sectors.
ESG/sustainability reporting regulations in the EU and Poland do not prohibit the use of AI for the purposes of reporting itself. However, given that the reporting regulations are new and only a limited number of reports drawn up in accordance with the recently adopted standards have thus far been published, the usability of AI in the drafting of reports may be limited by insufficient input data.
AI may, however, play a significant role in the analysing of vast amounts of data involving environmental metrics (such as carbon emissions, energy consumption, and waste management figures), social metrics (like employee diversity, labour practices, and community engagement) and governance metrics (board composition, executive compensation, and ethical business practices). In this field, AI may assist companies in identifying trends, assessing risks, and making informed decisions, thereby enhancing their sustainability efforts.
Based on market practice, the following best practices are usually adopted in organisations implementing AI solutions.
Jasna 26 Street
00-054 Warsaw
Poland
+48 22 608 70 00
+48 22 608 70 01
office@skslegal.pl www.skslegal.pl