Contributed By ICT Legal Consulting
The Italian regulatory framework on the protection of personal data and privacy is dictated by Regulation (EU) 2016/679 of the European Parliament and of the Council of 27 April 2016 on the protection of individuals with regard to the processing of personal data and on the free movement of such data, repealing Directive 95/46/EC (GDPR). To the extent that such protection is not mentioned by the GDPR, it is regulated by Legislative Decree No 196/2003 (the “Privacy Code”).
Further detailed rules are contained in Directive 2002/58/EC of the European Parliament and of the Council of 12 July 2002 concerning the processing of personal data and the protection of privacy in the electronic communications sector, as transposed into Italian law by the Privacy Code.
With reference to the protection of natural persons with regard to the processing of personal data by competent authorities for the purposes of the prevention, investigation, detection and prosecution of criminal offences or the execution of criminal penalties, the regulatory framework is instead governed by EU Directive 2016/680, transposed into the Italian legal system through Legislative Decree No 51/2018. Finally, other specific indications and/or interpretations are contained in the decisions, recommendations and guidelines issued by the national supervisory authorities and the European Data Protection Board (eg, in Italy, the requirements for system administrators).
Concerning the extraterritoriality application, Article 2 of the GPDR addresses the application to the processing of personal data of data subjects who are in the EU by a controller or processor not established in the EU, where the processing activities are related to:
Finally, as regards the interaction with artificial intelligence (AI) rules (Regulation (EU) 2024/1689 and Italian Law No 132/2025) and with cybersecurity rules (in particular, Italian Legislative Decree No 138/2024 transposing Directive (EU) 2022/2555, together with the determinations of the National Cybersecurity Agency in the NIS2 framework, Italian Law No 90/2024, and all implementing measures of Regulation (EU) 2022/2554), these instruments expressly defer to EU personal data protection law, leaving its content and scope unchanged and fully applicable.
General Principles Relating to Processing of Personal Data
According to Article 5(1) of the GDPR, personal data shall be:
In addition, the data controller shall be responsible for, and be able to demonstrate compliance with, such principles (“accountability”).
Lawfulness of Processing
Any processing of personal data must be based on at least one of the following legal bases provided for in Article 6(1) of the GDPR:
Data Subjects’ Rights
Data subjects have certain rights under the GDPR in order to allow them to have continuous and effective control over their personal data. In particular, data subjects have the right to:
Special Categories of Personal Data
According to Article 9 of the GDPR, processing of personal data revealing racial or ethnic origin, political opinions, religious or philosophical beliefs, or trade union membership, as well as the processing of genetic data, biometric data for the purpose of uniquely identifying a natural person, data concerning health, or data concerning a natural person's sex life or sexual orientation, is prohibited except in the following cases:
According to Article 2-septies of the Privacy Code, the Italian supervisory authority (Garante per la Protezione dei Dati Personali; GPDP) may also specify particular safeguards for processing genetic data, biometric data and data concerning health.
Personal Data Relating to Criminal Convictions and Offences
According to Articles 10 of the GDPR and 2-octies of the Privacy Code – as amended by Article 71(1)(g) of Law No 182/2025 – processing of personal data relating to criminal convictions and offences or related security measures is carried out only under the control of official authority or when the processing is authorised by the law providing for appropriate safeguards for the rights and freedoms of data subjects.
Data Relating to Minors
In accordance with Article 8 of the GDPR and Article 2-quinques of the Privacy Code, restrictions on the processing of minors' data concern their consent. In particular, minors who have reached the age of 14 may consent to the processing of their personal data in relation to the direct offer of information society services (below this age, consent must be given by those exercising parental responsibility). However, the data controller must draft the information and communications relating to the processing in particularly clear and simple, concise and comprehensive language that is easily accessible and understandable to the minor.
Life sciences or medtech companies can anonymise patient data for product development or scientific research only if the outcome is true anonymisation in the GDPR sense. That means that the data must be rendered irreversibly non-identifiable, taking into account all means reasonably likely to be used for re-identification. In healthcare, this is a high bar, as clinical datasets are rich and easily linkable; the GPDP’s approach is consistently strict, so mere removal of direct identifiers or standard pseudonymisation is not enough. If the company achieves genuine anonymisation, the dataset falls outside the GDPR and can be used for research and development without a GDPR legal basis.
Where full anonymisation is not realistically achievable, data remains personal (even if pseudonymised). Then, the company must rely on a GDPR Article 6 lawful basis and an Article 9(2) condition for health data, most often the scientific research ground combined with Article 89 safeguards. The Italian Privacy Code rules on research (including the updated Article 110) support this pathway by allowing certain research uses with enhanced safeguards when informing individuals is impossible or disproportionate, but they do not remove the need for a lawful basis and accountability.
The Code of Conduct for management software developers approved by the GPDP in October 2024 reinforces this compliance logic for software/tech vendors working with hospitals: it operationalises privacy-by-design/default, clarifies that vendors typically act as processors for installation/maintenance/support, and expects robust security and life cycle controls. In practice, it pushes medtech vendors to ensure that any re-use/anonymisation for R&D is clearly covered contractually with the healthcare provider and supported by appropriate technical measures.
Looking forwards, the European Health Data Space Regulation (EU) 2025/327 (EHDS) is a major change for these companies. It entered into force on 26 March 2025 and will apply in phases from 26 March 2027, with the secondary use regime (research/innovation/AI training, etc) becoming operational from 26 March 2029 and some data categories being phased in later. The EHDS will create a harmonised, permit-based route to access large volumes of electronic health data across the EU through national Health Data Access Bodies, typically in secure environments and often in anonymised/pseudonymised form. For life sciences and medtech firms, this should increase lawful access and legal certainty for research and development, but only under tighter governance, interoperability, auditability and anti-reidentification constraints.
With regard to the use of AI systems, it is essential to note that both EU (AI Act) and national legislation refer to the GDPR and the Privacy Code for the regulation of personal data protection profiles relating to the use of AI tools. In this context, the first thing to note is the transparency obligations, as Articles 13 and 14 of the GDPR require data subjects to be provided with information on the existence of automated decision-making, including profiling, and meaningful information about the logic involved, as well as the significance and the envisaged consequences of such processing for the data subject.
In addition, Article 22 of the GDPR states that the data subject shall have the right not to be subject to a decision based solely on automated processing, including profiling, which produces legal effects concerning them or similarly significantly affects them, except where the decision is:
In these cases, the data controller shall implement suitable measures to safeguard the data subject’s rights and freedoms and legitimate interests – at least the right to obtain human intervention on the part of the controller – to express their point of view and to contest the decision.
Lastly, it is likely that an AI tool is also subject to the Data Protection Impact Assessment requirement under Article 35 of the GDPR.
In the case of a personal data breach, the data controller shall, without undue delay and where feasible no later than 72 hours after having become aware of it, notify the personal data breach to the GPDP, unless the personal data breach is unlikely to result in a risk to the rights and freedoms of natural persons. Such notification should at least:
Where the notification to the authority is not made within 72 hours, it shall be accompanied by reasons for the delay. In addition, where and in so far as it is not possible to provide all the information at the same time, the information may be provided in phases without undue further delay.
When the personal data breach is likely to result in a high risk to the rights and freedoms of natural persons, in addition to the notification described above, the data controller shall also communicate the personal data breach to the data subject without undue delay, describing in clear and plain language the nature of the personal data breach and the same above-mentioned information. Such communication to the data subject shall not be required if:
Lastly, the controller shall document any personal data breaches, comprising the facts relating to the personal data breach, its effects and the remedial action taken.
Supervisory authorities have limited regulatory power, mainly through the adoption of guidelines and opinions interpreting legal provisions. However, supervisory authorities (in Italy, the GPDP) also have supervisory powers to monitor compliance with data protection legislation, and benefit from investigative powers that include, ex multis, the possibility of requesting information from data controllers and data processors or conducting on-site checks and inspections.
In this context, the supervisory authority may request access to the documentation adopted (privacy policy, consents, internal policies and procedures, records of processing activities, etc) and to systems and databases. The GPDP’s inspections may be triggered by the authority itself (based on an inspection plan adopted and published every six months, or following notification of a personal data breach) or by data subjects or other third parties (in the case of complaints or reports). Any decisions that are eventually adopted are published.
Data protection legislation may also be applied by the courts in the case of appeals lodged by individuals (particularly in the case of claims for damages or appeals against decisions of the supervisory authority).
As mentioned in 1.7 Regulators, GPDP inspections may be triggered by the authority itself (on the basis of an inspection plan adopted and published every six months, or following the notification of a personal data breach) or by data subjects or other third parties (in the case of complaints or reports).
Preliminary Investigation
In the event of a complaint by a data subject, the GPDP shall verify the correctness and completeness of the complaint and, if necessary, grant the complainant a period of time to amend it, normally not exceeding 15 days. In the event of a correct and complete complaint (or in the event of an investigation on its own accord, such as following the notification of a personal data breach), the GPDP shall start a preliminary investigation, during which the documentation received is examined and/or further information is requested from the data controller or data processor.
In that scenario, inspections may also be carried out, during which the entity subject to inspection may be assisted by its trusted advisers and reserve the right to produce the documentation that is not immediately available within a reasonable period (as a rule, not exceeding 30 days). A record of the activity carried out shall also be drawn up, with particular reference to the statements made and the documents acquired, and a copy shall be given to the subject under inspection.
Closing of the Preliminary Investigation and Archiving
At the end of the preliminary investigation, the competent department within the GPDP may conclude its examination of the complaint by archiving it, when:
In the case of a complaint, feedback is provided to the applicant, briefly stating the reasons why no action is being taken.
Initiation of Proceedings
If the matter is not dismissed following the preliminary investigation, the competent department shall initiate proceedings for the adoption of measures by the board of the GPDP, by means of its own communication to the data controller and/or data processor. The communication shall contain:
Right of Defence
The addressee of the notice may exercise the right of defence by submitting written statements and documents within 30 days from the date of notification of the communication, as well as a personal testimony regarding the facts of the notice, where requested.
The addressee of the notice may request a short extension by specifically and duly motivating the request. The extension shall normally not exceed 15 days, and may be granted according to proportionality criteria and criteria relating to the operational/dimensional characteristics of the addressees themselves and to the complexity of the matter under examination. The addressee of the notice may also request a hearing before the GPDP.
Failure to submit written counterarguments or a request for a hearing shall not prejudice the continuation of the proceedings.
Decision
Where necessary, the board of the GPDP, by its own resolution, shall adopt the corrective and sanctioning measures referred to in Article 58(2) of the GDPR (in the case of an administrative pecuniary sanction, the quantum is calculated on the basis of the criteria indicated by Article 83 of the GDPR). The decision is notified to the parties by the department, service or other organisational unit that has supervised the preliminary investigation.
Appeal Against Measures of the GPDP
Under penalty of inadmissibility, an appeal against the measures adopted by the GPDP must be lodged within 30 days from the date of communication of the decision or within 60 days if the appellant resides abroad, with the ordinary court of the place where the data controller resides, or with the court of the place of residence of the data subject. At the time of the appeal, it is also possible to request the court to suspend the enforceability of the contested decision.
The so-called “work ritual” applies to the judicial procedure, and the sentence that defines the judgment is not appealable before the judge and may prescribe the necessary measures and compensation for damages.
Over the past 24 months, Italian enforcement has concentrated on a few high-risk clusters:
A parallel trend is the growing privacy–competition convergence, with Italian Competition Authority (AGCM) scrutiny on consent and data-combination practices.
In recent years, data protection litigation in Italy has experienced significant growth, driven by an increasing awareness of rights among data subjects.
The most frequent disputes involve unlawful data processing, data breaches, and the improper use of personal data by companies and public administrations. Claims for compensation for privacy violations, particularly for non-material damages, are also on the rise.
In this context, the GPDP is playing a crucial role, imposing substantial fines that influence corporate strategies, seeking to stay ahead of other European authorities, and positioning itself as a leader on issues related to AI (for instance, the proceedings initiated against OpenAI and ChatGPT, which concluded with a sanction in December 2024) and employee monitoring, especially concerning the retention of metadata generated through employees’ use of email tools (ie, the decision against Regione Lombardia on metadata and logs relating to internet browsing).
The most significant recent decision for EU/Italian privacy litigation is the CJEU’s Deloitte judgment C-413/23 P (EDPS v SRB), 4 September 2025. The Court held that pseudonymised data is not automatically “personal data” for every actor: for a third-party recipient, the qualification depends on whether that recipient has reasonably likely and practically available means to re-identify individuals. If, in the recipient’s hands, identification is not realistically possible, the data may be treated as anonymous for that recipient, even though it remains personal data for the original controller who retains the re-identification key. The Court set aside the General Court’s approach because it had relied on an abstract presumption (pseudonymisation = personal data) without assessing Deloitte’s concrete re-identification capability.
The national legislation on personal data protection does not currently provide explicit regulation for collective redress mechanisms. However, data subjects may rely on the tools generally available under civil procedure law or those designed to protect their rights as consumers.
In Italy, the protection and governance of non-personal data is largely shaped by directly applicable EU regulations, rather than by autonomous national legislation. There is no standalone Italian “horizontal” law equivalent to the EU Data Act; instead, Italy relies on the EU data strategy instruments, complemented only by sectoral or administrative measures.
The starting point is the Free Flow of Non-Personal Data Regulation (Regulation (EU) 2018/1807), which establishes the principle of the free movement of non-personal data within the EU. It prohibits unjustified data-localisation requirements imposed by member states, and promotes portability and switching for professional users of data processing and cloud services. Importantly, for mixed datasets (where personal and non-personal data are intertwined), the Regulation operates alongside the GDPR: the GDPR applies to the personal component, while the free-flow rules apply to the non-personal component.
This baseline is expanded by the Data Governance Act (Regulation (EU) 2022/868), which provides a cross-sector EU framework to increase data availability and sharing, covering both personal and non-personal data. It regulates the re-use of certain protected data held by public sector bodies, introduces trusted roles such as data intermediation service providers, and supports voluntary mechanisms such as data altruism and the development of Common European Data Spaces. Italy has adopted implementing measures mainly to designate competent authorities and procedural arrangements, but the substantive regime remains EU-based.
Finally, the EU Data Act (Regulation (EU) 2023/2854) – applicable across the EU, including Italy, from 12 September 2025 – constitutes the central horizontal regime for access to and use of data generated by connected products and related services (ie, IoT ecosystems). It sets out rights and obligations for key actors such as data holders, users (B2C/B2B) and data recipients, and applies to both personal and non-personal data depending on the nature of the dataset, subject to GDPR compliance for any personal elements. A core pillar is the mandatory B2B/B2C data access and sharing framework, often under FRAND (fair, reasonable and non-discriminatory) terms, and another pillar addresses the cloud and edge layer by imposing rules on switching, interoperability and the removal of contractual/technical lock-ins in data processing services.
Beyond these EU Regulations, Italy has not enacted a separate Data Act-like statute. National activity is mainly confined to sector-specific governance (especially for public-administration data and cloud security classifications) and to procedural implementation where EU Acts require member state designation of authorities or enforcement structures.
These frameworks are designed to complement, not override, privacy and IP rules, so the interaction works in layers. When the data to be accessed or shared includes personal data (or is part of a mixed dataset), the GDPR remains fully controlling. The Data Act and the Data Governance Act do not create a new lawful basis on their own: any disclosure must still rely on an Article 6 GDPR ground already applicable to the data holder, and must respect all GDPR principles, including confidentiality, minimisation, purpose limitation, transparency and security. In practice, the user’s Data Act access right can be exercised only within what is lawful under the GDPR, and any recipient of personal data must comply with the GDPR as an autonomous controller or processor.
For non-personal data, access and sharing obligations under the Data Act/Data Governance Act are balanced against IP rights and trade secret protection. The general rule is that sharing goes ahead, but the data holder may and should impose proportionate safeguards – for example, confidentiality clauses, technical access controls, or limiting the dataset to what is strictly necessary – so that disclosure does not destroy trade secret value or infringe IP. A refusal to share is allowed only in exceptional cases where no reasonable protective measures could prevent a serious risk of economic harm.
Thus, conceptually, the system is a “dual compliance gate”: the GDPR determines whether and how personal data can move; trade secret/IP law determines under what protections sensitive non-personal data can be shared, with access remaining the default outcome.
As a baseline, the Free Flow Regulation guarantees that non-personal data can be stored and processed anywhere in the EU, banning unjustified localisation rules. It also promotes portability and switching for professional cloud users, mainly through industry codes of conduct, and confirms that mixed datasets are governed jointly by the GDPR (personal layer) and free-flow rules (non-personal layer).
The Data Governance Act does not impose broad private-sector access duties but sets trusted conditions for sharing. It regulates the re-use of certain protected public-sector data, creates a compliance regime for neutral data intermediation services, and enables voluntary “data altruism” channels.
The Data Act is the core horizontal regime. It gives users of connected products and related services a right to access the data they generate, and to require the data holder to share those data with a third party (data recipient). Data holders must design products/services so that data is accessible by default, and must share B2B data under FRAND terms, with additional safeguards for SMEs and controls on unfair contractual clauses. In parallel, the Act imposes strong switching, portability, interoperability and termination duties on providers of data processing services (cloud/edge), including removal of technical/contractual lock-ins and limits on switching/egress fees. Any flow involving personal data remains fully subject to GDPR requirements; the Data Act does not override them.
For organisations, the main action items are practical and contractual:
In Italy, enforcement of the EU non-personal data regime is shared among multiple authorities. AgID is the lead authority for the Data Governance Act, overseeing data intermediation services, data altruism registrations, and public sector data re-use; it also functions as the co-ordination hub. When data sharing involves personal or mixed datasets, AgID works with the GPDP to ensure GDPR compliance. Where access or FRAND terms create competition or unfair-contract concerns, the AGCM intervenes, and the ACN provides support in cybersecurity/cloud-security aspects.
For the Data Act, Italy is moving towards a multi-authority model: digital/data-governance functions close to AgID, technical/market-surveillance authorities for connected products, the GPDP for the personal data layer, and the AGCM for FRAND, unfair terms and anti-lock-in issues. The main trend is co-ordinated enforcement: Data Act cases are increasingly treated as “hybrid” matters requiring simultaneous privacy, competition and technical oversight, with regulators prioritising practical outcomes (IoT user access and cloud switching) over purely formal classifications.
In Italy, the general rule is set out by Article 122 of the Privacy Code (which transposes Directive 2002/58/EC), under which all cookies – and other similar tracking tools – other than those strictly necessary for the functioning of the website may be installed on the users’ devices only with their consent.
In this regard, as clarified by the Guidelines adopted by the Italian Supervisory Authority in 2021, it is essential that consent for the use of cookies is collected in compliance with the principles established by the GDPR. Accordingly, it must be preceded by a multi-layered notice (cookie banner) that provides information about the cookies and the related personal data processing, and that allows the user to freely accept or refuse the use of cookies as well as to change their decision at any time.
A specific consideration applies to analytical cookies, which may be treated as necessary cookies (and therefore not require consent) only when:
In Italy, the general rule is set out by Article 130(1;2) of the Privacy Code (which transposes Directive 2002/58/EC), under which commercial and promotional communications by email, fax, telephone and similar means of communication require the prior consent of the user (natural or legal person). However, Article 130(4) provides for an exception to the requirement of consent, allowing for the processing of the email address provided by the data subject in the context of the sale of a product or a service for the purpose of sending commercial communications aimed at the direct sale of products or services similar to those already purchased, provided that the data subject has been adequately informed and does not refuse such use, either initially or on the occasion of subsequent communications.
With specific regard to telephone marketing activities, Article 130(3-bis) provides that data controllers may lawfully contact all users who have not objected to receiving commercial communications by telephone by registering in the Register of Opposition. In this sense, pursuant to Law No 5/2018, users may enlist in the register in order to prevent subsequent communications and, at the same time, withdraw any consent previously given to the processing of their personal data for telephone marketing purposes. In fact, the data controller who intends to carry out telemarketing activities is required to consult the register at least every 15 days or, in any event, before the start of a new campaign.
On the other hand, online marketing may consist primarily of an activity carried out through the use of profiling and advertising cookies (see 4.1 Use of Cookies), or of behavioural advertising and targeting activities carried out through the use of external databases (especially those of social networks). In this second case, the jurisprudence of the CJEU and the interpretation provided by the European Data Protection Board (EDPB) in Guidelines 8/2020 clarify the need to carry out the activity on the basis of the prior consent of the data subject and, as a general rule, to reconstruct the privacy roles between the company and the social network as joint controllers of the processing to be regulated under Article 26 of the GDPR.
Processing carried out in the employment context is one of the sectors to which the GDPR defers to its regulation under national law, without prejudice to certain common guidelines and orientations first shared by WP29 and then by the EDPB, specifically regarding the vulnerable position of the data subject employee vis-à-vis the data controller employer (a situation that results in the presumption of the invalidity of any consents requested from the employee due to a lack of freedom).
Managing the Selection Process and the Employment Relationship
In these phases, the employer’s activities must respect – more than ever – the principle of minimisation, ensuring that only personal data that is essential for the performance of work duties and that, to a large extent, is governed by labour law provisions (eg, Article 8 of Law No 300/1970 or Legislative Decree No 81/2008) is requested from the candidate or employee.
Remote Monitoring of Workers
Without prejudice to a general prohibition on the use of instruments (also based on AI) to monitor employee activities, this case is governed by Article 4 of Law No 300/1970, which legitimises the use of such tools solely for organisational purposes and the protection of company assets (eg, cybersecurity purposes). In this case, without prejudice to instruments that are essential and prearranged for the performance of work duties, the use of instruments for remote monitoring is permitted only if doing so is:
In these cases, the employee data subject will have to be provided with additional and detailed information on what is normally provided for under Articles 13 and 14 of the GDPR; this can be done by adopting an internal regulation on the use of IT tools, for example, which also informs employees of the possible controls and their purposes.
However, although the agreement with trade union representatives or administrative authorisation is sufficient to legitimise the activity from the point of view of labour law, this does not exempt the employer from complying with the principles on the protection of personal data (eg, the principle of minimisation). In this sense, unencrypted or clear monitoring of the URLs surfed by employees is unlawful because, in terms of security purposes, the same results can be achieved by implementing filters that inhibit the surfing of potentially risky websites. On this point, see also the Guidelines adopted by the GPDP on 1 March 2007 or the Guidelines on the metadata collected through the use of email accounts.
Whistle-Blowing and the Transparency Decree
The national legislation on whistle-blowing was updated to transpose Directive (EU) 2019/1937 through Legislative Decree No 2023/24, which made discipline uniform between the private and public sectors. With regard to the protection of personal data, the general principles dictated by the GDPR remain valid, concerning the obligations to set up reporting and management processes in compliance with the principles of privacy by default and by design, and with the need to ensure the confidentiality of the reporter (resulting in the inadequacy, for instance, of the email channel), carry out a Data Protection Impact Assessment (DPIA) on the processing, train and instruct the people who access the data and manage the reports, etc.
Further obligations (mainly informative) are also imposed by Legislative Decree No 104/2022 (the so-called “Transparency Decree”), which prescribes the need to carry out a DPIA and to provide additional information to data subjects in the event of “the use of automated decision-making or monitoring systems designed to provide indications relevant to the recruitment or assignment, management or termination of the employment relationship, the assignment of tasks or duties, as well as indications affecting the monitoring, assessment, performance and fulfilment of contractual obligations of workers”.
The value of personal data and consent databases as a corporate asset is often underestimated in corporate transactions. In this context, with regards to the sector in question, the main activity may consist of verifying the lawfulness and correctness of the processing of personal data that makes up a company's databases; this can be done by verifying the correctness and completeness of the information that the data controller had to provide to the data subjects pursuant to Articles 13 and 14 of the GDPR, and by examining the evidence of compliance with this information notice obligation.
Furthermore, where the processing of personal data is based on consent (eg, in the case of processing for promotional purposes or in the context of scientific research), it is essential to verify the correctness and ability to prove the consents collected from the data subjects and the effective capacity of the systems to receive any requests for withdrawal and/or opposition.
Following the M&A transaction, it is also essential that data subjects receive a communication informing them of the change in the identity of the data controller, as well as any other changes relating to the processing of personal data (new contact channels for the data controller and the Data Protection Officer, new methods for exercising rights, any transfer of data to third countries, etc). This communication should be provided upon first contact with the data subject or within 30 days.
European data protection legislation requires that any transfer of personal data that is undergoing processing or is intended for processing after transfer to a third country or to an international organisation (including for onward transfers of personal data from the third country or an international organisation to another third country or to another international organisation) shall take place only if the level of protection of natural persons guaranteed by the GDPR is not undermined.
This therefore requires an examination of the legal provisions applicable to the third country or international organisation in order to understand the actual level of protection of personal data, taking into account the elements specified in Article 45(2) of the GDPR. This analysis is carried out by the European Commission when it adopts the adequacy decisions referred to in Article 45 of the GDPR (decisions legitimising the transfer of personal data to the country or organisation benefiting from it).
In the absence of an adequacy decision, as clarified by the CJEU in its judgment of 16 July 2020 (the Schrems II judgment), this assessment is instead the responsibility of the data controller or data processor who is intending to export the personal data. In such a case, where the law in force in the third country or applicable to the international organisation does not guarantee an adequate level of protection of personal data, the transfer may only be carried out subject to the adoption of additional security measures suitable to mitigate the risks to the rights and freedoms of the data subjects (eg, encryption of the data prior to the transfer in order to exclusively share encrypted data).
With reference to non-personal data, see 5.2 Government Notifications and Approvals and 5.3 Data Localisation Requirements.
With reference to personal data, notification to the supervisory authority is only required in the case of transfers pursuant to Article 49(1)(2) of the GDPR. This is the case when no other means can be used to legitimise the transfer, and requires that the transfer:
With reference to non-personal data, Article 32 of the Data Act works as a safeguard against third-country authorities trying to obtain non-personal data stored in the EU from cloud/data-processing providers. A foreign court order or administrative decision is enforceable in the EU only if it is based on an international agreement with the EU or a member state (for example, a mutual legal assistance treaty). If no such agreement exists and compliance would conflict with EU or national law, the provider may grant access only where the third-country legal system offers solid rule-of-law guarantees:
To assess this, the provider can ask for an opinion from the national authority competent for international legal co-operation, especially where the request concerns trade secrets, commercially sensitive data, IP-protected material, or risks of re-identification; that authority may consult the Commission. A similar check applies if national security or defence interests may be involved. If the authority does not respond within one month, or concludes that the safeguards are not met, the provider is entitled to refuse the foreign request.
With reference to personal data, European legislation on the protection of personal data does not provide for any obligation to store data within a specific member state or the EEA, aiming, on the contrary, to regulate and facilitate the free movement of such data. In the case of transfers of data to third countries, however, the provisions of Chapter V of the GDPR apply in order to guarantee an adequate level of protection of personal data (see 5.1 Restrictions on International Data Transfers).
With reference to non-personal data, Article 32 of the Data Act states that the providers of data processing services shall take all adequate technical, organisational and legal measures, including contracts, in order to prevent international and third-country governmental access and transfer of non-personal data held in the EU where such transfer or access would create a conflict with Union law or with the national law of the relevant member state.
There are no “blocking” statutes in European legislation in addition to those described in the previous sections concerning the transfer of data outside the EEA.
The most recent update was in 2024, when the European Commission concluded the review process of 11 adequacy decisions regarding the transfer of personal data, confirming their validity. In this regard, the decisions remain effective, and it is permissible to continue the free transfer of personal data to Andorra, Argentina, Canada, the Faroe Islands, Guernsey, the Isle of Man, Israel, Jersey, New Zealand, Switzerland and Uruguay.
ICT Legal Consulting
Via Borgonuovo 12
20121 Milan
Italy
+39 0284 2471 94
+39 0270 0512 101
info.legal@ictlc.com www.ictlegalconsulting.com