Data Protection and Privacy 2020 covers 24 jurisdictions. This new guide includes outlines of national regimes, fundamental laws, sectoral or special issues, workplace privacy, enforcement and litigation and emerging issues.
Last Updated: March 09, 2020
Digital Governance: Regulating Privacy and Data Protection for Emerging Technologies
Citizens, consumers and companies could use more guidance than they’re currently getting on the governance and accountability principles appropriate to managing personal and sensitive data, and the deployment of human-centric digital technologies. The world’s regulation of privacy and data protection as well as enforcement related to disinformation, micro-targeting and other profiling, behavioural and location tracking, artificial intelligence, facial recognition technology and the collection of data from ubiquitous sensors and connected devices is evolving with insufficient analysis and co-ordination. Emotion and presumption appear to be dictating too much policy development in state after state and country after country. More attention must be focused on how emerging technologies actually hurt people and critical societal functions – and how to abate those harms – while honestly addressing and balancing the trade-offs.
Given that the USA has not yet enacted comprehensive federal legislation, maybe the time is right for it to start leading with a fresh set of insights. Congress could do this immediately by adopting legislation that would establish a Digital Accountability and Privacy Co-ordinator in the Executive Office of the President and empowering that official to develop information technology policy in the White House and federal agencies. The new Co-ordinator should also be authorised to convene state and international regulatory bodies in the interests of advancing, rationalising and harmonising the world’s approach to regulating the least territorial commodity in the history of the globe – data.
If we do not create a clear consensus on governing principles soon, other nations that may not equally share our values – if not machines themselves – will certainly do it for us before too long. A few thoughts are offered below on the relevant principles for digital policy development and co-ordination.
Policymakers ought to spend more time defining what “data” risks warrant government regulation and considering whether and how cost-benefit analysis should be applied to a fundamental right like privacy. Rigorous analysis is especially crucial when intangible data-related or digital harms are at stake. Systematically characterising the harms that warrant prevention, abatement, deterrence and punishment comprises and defines the “benefits” of privacy and digital regulation. The “costs” of such regulation are embodied in the price of trading off other societal interests if data and digital technology are either “over-regulated" or enforcement resources are misdirected toward illusory, rather than real, risks.
The right balance can be achieved by identifying the actual harms and risks of abusive data practices that warrant prohibition or restriction. Examples of concrete harms would obviously include identity theft; significant reputational embarrassment; revelation of sensitive private facts; bias and discrimination; lost economic opportunity; adversely affected significant interests like legal outcomes and sentencing, housing, insurance, etc.
Less obvious but nonetheless real risks would include manipulation of thought through micro-targeting and false information and loss of human agency or opportunity through profiling and micro-targeting; automated decision-making; or technologies relying on AI, facial recognition, predictive profiling, social scoring, etc.
On the other hand, the relevant trade-offs and costs of potential over-regulation that society needs to consider include handicapping the ability of governments to protect their citizens’ personal safety and physical security; encroachments on freedom of speech and the right to receive information; impaired technological innovation; limits on economic growth; constrained organisational flexibility; diminished consumer choice; reduced personal convenience; etc.
Given the universal acknowledgement among Western democracies that privacy is a fundamental human right, it bears emphasis that there is both a legal obligation and a practical necessity to balance privacy and data protection against other fundamental rights and important interests of society. The European Union’s principle of proportionality is essentially the embodiment of this “balance.”
Lest there be any doubt about the necessity to balance the fundamental human rights of privacy and data protection against other considerations, the legal authorities excerpted below should remove it:
Recital (4) of the EU’s 2016 General Data Protection Regulation (GDPR), states:
"The processing of personal data should be designed to serve mankind. The right to the protection of personal data is not an absolute right; it must be considered in relation to its function in society and be balanced against other fundamental rights, in accordance with the principle of proportionality. This Regulation respects all fundamental rights and observes the freedoms and principles recognised in the Charter as enshrined in the Treaties, in particular the respect for private and family life, home and communications, the protection of personal data, freedom of thought, conscience and religion, freedom of expression and information, freedom to conduct a business, the right to an effective remedy and to a fair trial, and cultural, religious and linguistic diversity."
In applying GDPR in Google v CNIL (September 2019) (Right to Be Forgotten), the CJEU stated:
"The right to the protection of personal data is not an absolute right; it must be considered in relation to its function in society and be balanced against other fundamental rights, in accordance with the principle of proportionality. ... in particular the respect for private and family life, … the protection of personal data, freedom of thought, conscience and religion, freedom of expression and information [and] freedom to conduct a business … The balance between the right to privacy and the protection of personal data, on the one hand, and the freedom of information of internet users, on the other, is likely to vary significantly around the world."
In establishing and reorganising the Privacy and Civil Liberties Oversight Board (PCLOB) in 2004 and 2007, the US Congress stated (in 42 USC 2000ee) that:
“in conducting the war on terrorism, the Government may need additional powers and may need to enhance the use of its existing powers [and] This shift of power and authority to the Government calls for an enhanced system of checks and balances to protect the precious liberties that are vital to our way of life and to ensure that the … choice between security and liberty is a false choice, as nothing is more likely to endanger America’s liberties than the success of a terrorist attack at home. Our history has shown us that insecurity threatens liberty. Yet, if our liberties are curtailed, we lose the values that we are struggling to defend. … [And thus the purpose of PCLOB is to] analyze and review actions the executive branch takes to protect the Nation from terrorism, ensuring that the need for such [national security] actions is balanced with the need to protect privacy and civil liberties…”
Under the Federal Trade Commission Act, including with respect to privacy protections for consumers, in 1994, Congress required the FTC to apply cost-benefit analysis before prohibiting business practices as “unfair” (in 15 USC 45(n)):
"The Commission shall have no authority … to declare unlawful an act or practice on the grounds that such act or practice is unfair unless the act or practice causes or is likely to cause substantial injury to consumers which is not reasonably avoidable by consumers themselves and not outweighed by countervailing benefits to consumers or to competition. …"
In providing flexibility and protections for online platforms in 1996, Congress stated its policy to minimise government regulation of the internet to promote economic growth and innovation (in 47 USC 230):
"The rapidly developing array of Internet and other interactive computer services available to individual Americans represent an extraordinary advance in the availability of educational and informational resources to our citizens. … The Internet and other interactive computer services have flourished, to the benefit of all Americans, with a minimum of government regulation… Increasingly Americans are relying on interactive media for a variety of political, educational, cultural, and entertainment services… It is the policy of the United States—… to promote the continued development of the Internet and other interactive computer services and other interactive media;… to preserve the vibrant and competitive free market that presently exists for the Internet and other interactive computer services, unfettered by Federal or State regulation…"
In addressing the use of personal data for political campaigns in March 2019, the European Data Protection Board stated:
"Engaging with voters is inherent to the democratic process. It allows the preparation of political programmes, enables citizens to influence politics and the development of campaigns in line with citizens expectations. Political parties, political coalitions and candidates increasingly rely on personal data and sophisticated profiling techniques to monitor and target voters and opinion leaders. In practice, individuals receive highly personalised messages and information, especially on social media platforms, on the basis of personal interests, lifestyle habits and values. Predictive tools are used to classify or profile people’s personality traits, characteristics, mood and other points of leverage to a large extent, allowing assumptions to be made about deep personality traits, including political views and other special categories of data. The extension of such data processing techniques to political purposes poses serious risks, not only to the rights to privacy and to data protection, but also to trust in the integrity of the democratic process. The Cambridge Analytica revelations illustrated how a potential infringement of the right to protection of personal data could affect other fundamental rights, such as freedom of expression and freedom to hold opinions and the possibility to think freely without manipulation."
Accordingly, thoughtful digital policymakers should consider the following factors.
All in all, regulating privacy, data protection and digital technology is considerably more complex than one would gather from present legislation and regulation. We can hope, however, that enlightened new leadership somewhere in the world will help lead to more meaningful and protective digital policy soon.