Food prices near all-time highs

Considering how much abuse and plunder that the ruling class has inflicted upon the working class, it's amazing that there hasn't been a widespread revolt when you think about it. Sky-high housing, rents, and energy has put the poor under a great deal of pressure, but it's rising food prices that historically are the trigger.

fao.jpg

A global index released on Thursday by the United Nations Food and Agriculture Organization showed food prices in January climbed to their highest level since 2011, when skyrocketing costs contributed to political uprisings in Egypt and Libya. The price of meat, dairy and cereals trended upward from December, while the price of oils reached the highest level since the index’s tracking began in 1990.

Maurice Obstfeld, a senior fellow at the Peterson Institute for International Economics who was formerly chief economist at the International Monetary Fund, said that food price increases would strain incomes in poorer countries, especially in some parts of Latin America and Africa, where some people may spend up to 50 or 60 percent of their income on food.

He said that it wasn’t “much of an exaggeration” to say the world was approaching a global food crisis, and that slower growth, high unemployment and stressed budgets from governments that have spent heavily to combat the pandemic had created “a perfect storm of adverse circumstances.”

fao1.jpg

The chairman of Britain's biggest supermarket has warned "the worst is yet to come" on rising food prices.
Hunger was already rising and widespread.

The World Bank conducted rapid phone surveys in 72 countries and found that a significant number of people are running out of food or reducing their consumption. Between 720 million and 811 million people in the world went hungry in 2020, which the World Bank estimated to be around 118 million more people than in 2019.

The Department of Agriculture (USDA) estimates about 10 percent of the U.S. population is food insecure, defined as households that obtain enough food to avoid substantially disrupting their eating patterns or reducing food intake by using different coping strategies, like eating less varied diets, using federal food assistance programs and getting food from community food pantries.

In 2020, the USDA estimated household food insecurity affected about 15 percent of households with children, affecting about 6 million kids.

A Gallup poll found that most Americans expect inflation to stick around over the next six months, with 79 percent predicting inflation will go up and 50 percent saying they expect it to go up “a lot.”

So what's causing all of this? Supply chain issues? Maybe a bit.
But mostly it's because of corporate greed.

Food companies do face legitimate increased costs and unique shortages, but these aren’t eating into their profits as economists might expect. In fact, the largest publicly traded companies have never had higher profit margins. Such record earnings suggest that food companies have sufficient market power to pass all their higher costs, and then some, onto consumers. Basic economic theory tells us that when a business charges too much, competitors will offer lower prices, take sales, and erode excessive profit. Sustained, exceptional corporate profits raise the question: how much are food companies really competing?
...
Food production has consolidated dramatically since the 1970’s after changes in antitrust policy allowed more companies to buy up their competitors. Depending on who you ask, antitrust practitioners say markets are “oligopolistic” or dangerously concentrated when the top four firms control 40% to 50% of the market, or more. Higher levels of concentration give businesses more power to set prices and increase the likelihood of price-fixing or market manipulation. Today, the top four corporations control more than 60% of the U.S. market for pork, coffee, cookies, beer, and bread. In beef processing, baby food, pasta, and soda the top four companies control more than 80% of the U.S. market.

The serfs will tolerate an incredible amount of suffering, oppression and exploitation. Want to steal all of the land and jack up rents to nose-bleed levels? The serfs will often tolerate it, but hunger has a way of making the most docile serfs rise up against their ruling elites.
As we've all seen, corporate greed is completely without limits. Historically governments have reluctantly stepped in to curb the rapaciousness of corporations when they threaten the existing power structure. However, the governments of today are 100% captured by the corporate class and are completely unable to curb corporate abuse, even to save themselves. So we are rapidly approaching a point of conflict.

Share
up
20 users have voted.

Comments

The Liberal Moonbat's picture

up
7 users have voted.

In the Land of the Blind, the One-Eyed Man is declared mentally ill for describing colors.

Yes Virginia, there is a Global Banking Conspiracy!

Where I live here in Ann Arbor MI, two of the three full-size grocery stores we regularly visit were decimated on canned goods and frozen foods last weekend. I haven’t gotten to the third store or the coop market I also frequent yet.

I asked an employee at one of the stores and she told me, “The trucks haven’t shown up all week.” I’m not sure what to make of that.

up
14 users have voted.

then this is all predictable. Since the oil embargo in the 70's there are big profits in shortages, real or manufactured. All possible due to consolidation, vertical integration, profit extraction and the Supreme Court. The republicans blame the democrats, and the democrats blame "the left". It's all our fault.

It used to be unions were the whipping boy, now it's us. We want stuff, like decent wages and health care, and are ungrateful enough to not be satisfied with multi-million dollar jet fighters and credit default swaps. If you're poor it's your fault, and if poverty is rampant then there will be help. For a narrow tailored segment deemed exceptionally deserving, served by private/government programs that consumes 90 cents out of every dollar in administration for every 10 cents actually given to the poor.

At least with local charities you can see where the money goes. They're pretty basic, if you're hungry, here's food. If you're cold, here's shelter. There was that song about war, what is it good for? Same could be said for government. Cause it ain't much good for us.

up
16 users have voted.

Whatever name it goes by, during the Pandemic, many states like mine increased the amount that people were allowed to receive. Everyone was given the maximum possible benefit for approximately 9 months to a year depending on your circumstances.

I was receiving $200.00 a month. Now that this is over, the President has increased the usual allotment by 25%. Now I am receiving $20.00 a month, which is about the equivalent of 3 gallons of milk and 3 loaves of bread.

up
10 users have voted.
The Liberal Moonbat's picture

@Enchantress ...
"A quart of wheat for a denarius, and three quarts of barley for a denarius, and do not harm the oil and wine."

Golly, all this and a pandemic, too!

And serious sabre-rattling in Eastern Europe...and near China...and genocide in Yemen...

...and...and...and...

up
9 users have voted.

In the Land of the Blind, the One-Eyed Man is declared mentally ill for describing colors.

Yes Virginia, there is a Global Banking Conspiracy!

Lookout's picture

One reason I harp on growing a garden.

up
8 users have voted.

“Until justice rolls down like water and righteousness like a mighty stream.”

@Lookout hi, I did get a Territorial Seed catalog and it is fun thinking about it, but sheesh the city here still has a 25% penalty per month, I can't afford to grow no mo. Newsom declared drought in January I think? It is really dry, and already starting to warm up. "Bud break" is how grape growers mark time, and I bet it will happen this week or next, though nights are still cold not freezing. It will be in the news when it happens I guess. Bring it on. go spring

Peace and Love

up
6 users have voted.
zed2's picture

@eyo @eyo

Called XTandi.

4 Times what they charge in the other richest countries. And we the people, our money paid to discover it. Our people need to speak up!

up
7 users have voted.

@zed2
That would wipe out the savings of most people in a year or two.

up
4 users have voted.
Lookout's picture

@eyo

water costs? My thought was many people have no yard, patio, or balcony so are limited, but water never crossed my mind here where we got 90" of rain last year.

Well let me know about if a new location comes about, and we'll connect.

up
8 users have voted.

“Until justice rolls down like water and righteousness like a mighty stream.”

usefewersyllables's picture

I'm sure our dem elected officials will get right on that.

After all, they've been telling us for years that that's what they are there for: to keep those nasty mean horrible republican assholes from allowing us to be priced out of existence. And every couple of years, we go to the polls and tell them "You're doing a heckuva job, Brownie...".

No more votes for dems from this voter. I'd prefer the overtly nasty mean horrible asshole to the smug sociopathic hypocrite with all the excuses. And I'd prefer someone not already owned in fee simple far above either of those. Fuck the dem party: they stand for nothing, other than their own graft.

up
13 users have voted.

Twice bitten, permanently shy.

zed2's picture

@usefewersyllables What makes you think tey will do anything other than parrot what drug companies tell them to say. I know my states Senator Robert Menendez does.

He is 100% owned by these drug companies, its always seemed to me. Its a shame Latin Americans arent smarter when it comes to the things that effect their lives a lot like drug prices.

I would be so much better off financially if I hadnt had to spend so much money out of pocket $300 a month for decades, on PRILOSEC. Thank God for generic drugs (which the big drug company donors just hate if they had their ways there would be no generic drugs) or I would still be paying an arm and a leg for it. I voted for pelosi dozens of times, and now I really really regret it. We need a complete house cleaning of these people who shamelessly sell us out. Kick them out!

up
4 users have voted.
zed2's picture

I think whats happening is an attempt by TPTB to mount a preemptive strike on humanity and especially the concept of "democracy" to capture and mold a dissolution of representative democracy, (except for global governance institutions that could hardly be called "democratic" as they are supranational and individual humans have no standing in them.) Whats being engineered is creation of a global, semi-totalitarian "digital Welfare State" fully controlled by corporations. All the evidence points there especially things like the large scale buying up of farms and the creation of situations that cause large numbers of farms and owner operated businesses to fail.

This is the beginning of a recent UN report on Extreme Poverty and Human Rights.

Report of the Special rapporteur on extreme poverty and human rights**
Note by the Secretary-General
The Secretary-General has the honour to transmit to the General Assembly the report of the Special Rapporteur on extreme poverty and human rights, Philip Alston, submitted in accordance with Human Rights Council resolution 35/19.

The digital welfare state is either already a reality or is emerging in many countries across the globe. In these states, systems of social protection and assistance are increasingly driven by digital data and technologies that are used to automate, predict, identify, surveil, detect, target and punish. This report acknowledges the irresistible attractions for governments to move in this direction, but warns that there is a grave risk of stumbling zombie-like into a digital welfare dystopia. It argues that Big Tech operates in an almost human rights free-zone, and that this is especially problematic when the private sector is taking a leading role in designing, constructing, and even operating significant parts of the digital welfare state. The report recommends that instead of obsessing about fraud, cost savings, sanctions, and market-driven definitions of efficiency, the starting point should be on how welfare budgets could be transformed through technology to ensure a higher standard of living for the vulnerable and disadvantaged.

I. Introduction 3
II. Uses of digital technologies in the welfare state 5
A. Identity verification 5
B. Eligibility assessment 7
C. Welfare benefit calculation and payments 8
D. Fraud prevention and detection 8
E. Risk scoring and need classification 9
F. Communication between welfare authorities and beneficiaries 9
III. Making digital technologies work for social protection 10
A. Taking human rights seriously and regulating accordingly 11
B. Ensuring legality and transparency 12
C. Promoting digital equality 13
D. Protecting economic and social rights in the digital welfare state 14
E. Protecting civil and political rights in the digital welfare state 15
F. Resisting the inevitability of a digital only future 17
G. The role of the private sector 18
H. Accountability mechanisms 18
IV. Conclusions 19
I. Introduction
The era of digital governance is upon us. In high and middle income countries, electronic voting, technology-driven surveillance and control including through facial recognition programs, algorithm-based predictive policing, the digitization of justice and immigration systems, online submission of tax returns and payments, and many other forms of electronic interactions between citizens and different levels of government are becoming the norm. And in lower income countries, national systems of biometric identification are laying the foundations for comparable developments, especially in systems to provide social protection, or ‘welfare’, to use a shorthand term.
Invariably, improved welfare provision, along with enhanced security, is one of the principal goals invoked to justify the deep societal transformations and vast expenditures that are involved in moving the entire population of a country not just on to a national unique biometric identity card system but on to linked centralized systems providing a wide array of government services and the provision of goods ranging from food and education to health care and special services for the ageing or those with disabilities.
The result is the emergence of the ‘digital welfare state’ in many countries across the globe. In these states, systems of social protection and assistance are increasingly driven by digital data and technologies that are used to automate, predict, identify, surveil, detect, target and punish. The process is commonly referred to as ‘digital transformation’, but this somewhat neutral term should not be permitted to conceal the revolutionary, politically-driven, character of many such innovations. Commentators have predicted “a future in which government agencies could effectively make law by robot”, and it is clear that new forms of governance are emerging which rely significantly on the processing of vast quantities of digital data from all available sources, use predictive analytics to foresee risk, automate decision-making, and remove discretion from human decision-makers. In such a world, citizens become ever more visible to their governments, but not the other way around.
Welfare is an attractive entry point not just because it takes up a major share of the national budget or affects such a large proportion of the population but because digitization can be presented as an essentially benign initiative. Thus, for example, the United Kingdom’s digital strategy proclaims that it will “transform the relationship between citizens and the state”, thus “putting more power in the hands of citizens and being more responsive to their needs.” In India, the core values of the Unique Identification Authority of India include: facilitating good governance, integrity, inclusive nation building, a collaborative approach, excellence in services, and transparency and openness.
In other words, the embrace of the digital welfare state is presented as an altruistic and noble enterprise designed to ensure that citizens benefit from new technologies, experience more efficient government, and enjoy higher levels of well-being. Often, however, the digitization of welfare systems has been accompanied by deep reductions in the overall welfare budget, a narrowing of the beneficiary pool, the elimination of some services, the introduction of demanding and intrusive forms of conditionality, the pursuit of behavioural modification goals, the imposition of stronger sanctions regimes, and a complete reversal of the traditional notion that the state should be accountable to the individual.
These other outcomes are promoted in the name of efficiency, targeting, incentivizing work, rooting out fraud, strengthening responsibility, encouraging individual autonomy, and responding to the imperatives of fiscal consolidation. Through the invocation of what are often ideologically-charged terms, neoliberal economic policies are seamlessly blended into what are presented as cutting edge welfare reforms, which in turn are often facilitated, justified, and shielded by new digital technologies. Although the latter are presented as being ‘scientific’ and neutral, they can reflect values and assumptions that are far removed from, and may be antithetical to, the principles of human rights. In addition, because of the relative deprivation and powerlessness of many welfare recipients, conditions, demands and forms of intrusiveness are imposed that would never have been accepted if they had instead been piloted in programs applicable to better-off members of the community.
Despite the enormous stakes involved not just for millions of individuals but for societies as a whole, these issues have, with a few notable exceptions, garnered remarkably little attention. The mainstream tech community has been guided by official pre-occupations with efficiency, budget-savings, and fraud detection. The welfare community has tended to see the technological dimensions as separate from the policy developments, rather than as being integrally linked. And those in the human rights community concerned with technology have understandably been focused instead on concerns such as the emergence of the surveillance state, the potentially fatal undermining of privacy, the highly discriminatory impact of many algorithms, and the consequences of the emerging regime of surveillance capitalism.
But the threat of a digital dystopia is especially significant in relation to the emerging digital welfare state. This report aims to redress the neglect of these issues to date by providing a systematic account of the ways in which digital technologies are used in the welfare state and of their implications for human rights. It concludes by calling for the regulation of digital technologies, including artificial intelligence, to ensure compliance with human rights, and for a rethinking of the positive ways in which the digital welfare state could be a force for the achievement of vastly improved systems of social protection.
The report builds in part on country reports by the Special Rapporteur on visits to the United States in 2017 and the United Kingdom in 2018 which drew attention to the increasing use of digital technologies in social protection systems. In preparing the report, the Special Rapporteur consulted with representatives of various digital rights groups, leading scholars, and other stakeholders, first in a meeting hosted by the Digital Freedom Fund in Berlin in February 2019, and then at another sponsored by the Center for Information Technology Policy at Princeton University in April 2019. In addition, a formal call for contributions resulted in some

(continued in my next post) This is the unedited first version.

up
1 user has voted.
zed2's picture

@zed2

Food security is the ultimate human right. Politics and politicians are to the starving, an unattainable unaffordable insulting and irrelevant luxury people cant even conceptualize if they dont have food, their children are dying, crying for milk and bodies are swollen from protein malnutrition.. History teaches us that starvation is the legacy of bad government. Famine is how a society treats those it deems unworthy of life. Increased climate variability and ENSO makes famines cyclic and unavoidable. The world has to set up means of providing people in drought afflicted areas food. The cult of the Invisible Hand that rules us views speculation on food during drought periods as desirable and legitimate. Hoever in a number of famines in recent years this dogma has proven to be hugely broken and wrong. Still it seems to have been accepted as truth during the huge Bengal Famine of 1945, when at least three million Indians died of starvation while citizens of the British Empire, arguably still the richest and most powerful nation on Earth. Here is the first, incomplete part of the UN special report by Philip Alston. Please try to read the full version, the footnotes are valuable and contain many links to must read works, (at least I consider them to be essential reading).

---------------------------------------

The report builds in part on country reports by the Special Rapporteur on visits to the United States in 2017 and the United Kingdom in 2018 which drew attention to the increasing use of digital technologies in social protection systems. In preparing the report, the Special Rapporteur consulted with representatives of various digital rights groups, leading scholars, and other stakeholders, first in a meeting hosted by the Digital Freedom Fund in Berlin in February 2019, and then at another sponsored by the Center for Information Technology Policy at Princeton University in April 2019. In addition, a formal call for contributions resulted in some 60 submissions from 34 countries, including 22 governments, as well as international and national civil society organizations, National Human Rights Institutions, academics, and individuals. While it is impossible to do justice to these rich and detailed submissions in such a necessarily brief report, the Special Rapporteur has made them electronically available and will continue analyzing them in the context of ongoing work on the digital welfare state.
II. Uses of digital technologies in the welfare state
From the many submissions received, and on the basis of various case studies addressed in the literature, it is possible to distinguish various ways, and different stages in the welfare context, in which digital innovation has been most prominently used.
A. Identity verification
Establishing every person’s legal identity, including through birth registration, by the year 2030 is the aim of Sustainable Development Goal (SDG) 16.9. A verifiable identity is essential for applying for benefits, establishing entitlement, receiving benefits, and appealing against denial. For the Government or other provider, it avoids duplication and fraud, it facilitates accurate targeting, and it enhances efficiency. Traditionally, paper and/or plastic documents have been used in forms such as birth certificates, identity cards and passports. These systems function reasonably well in most of the Global North, although 21 million adults in the United States do not have government-issued photo ID. In the Global South, 502 million people in sub-Saharan Africa and 357 million in South Asia lack official identification. In Liberia, for example, birth registration stands at only 5 per cent and national identity cards were not introduced until 2015.
In response, the World Bank, regional development organizations and bilateral donors have launched new programs to promote access to identity documents. In particular, the World Bank’s ‘ID4D’ (Identification for Development) campaign has focused heavily on promoting digital technologies as the key solution. This is explicitly stated in the ‘Principles on Identification for Sustainable Development’, which have been widely endorsed, including by MasterCard.
The Principles acknowledge that there are both pros and cons involved. On the positive side, it is claimed that digital technology can “create huge savings for citizens, governments, and businesses by reducing transaction costs, increasing efficiency, and driving innovation in service delivery, particularly to the poorest and most disadvantaged groups in society”. Digital identity systems can also “improve governance, boost financial inclusion, reduce gender inequalities by empowering women and girls, and increase access to health services and social safety nets for the poor”.
But in addition to this impressive and by now familiar sales pitch, the Principles, and similar documents, also recognize possible downsides ranging from political backlash to privacy and (cyber)security concerns. Solutions for dealing with those risks are often either technological or take the form of soft law norms. Thus a comparable United States Agency for International Development (USAID) document calls for “open source solutions” and developing good “practices for data privacy” to resolve the relevant problems. While the World Bank Principles refer to isolated human rights principles such as Article 7 of the Convention on the Rights of the Child, they rely primarily on the need to create an interoperable system using open standards, and protecting ‘privacy through system design’.
The world’s largest biometric identification system is Aadhaar in India, which is a 12-digit unique identifying number issued to Indian residents. It contains both demographic and biometric information, including an iris scan, a photograph and fingerprints. It is used to verify the identity of recipients of benefits and subsidies and is now mandatory to access those social rights. It was first introduced in 2009 and now covers more than 1.2 billion people, and has been enthusiastically endorsed by the international development community. The World Bank has praised it for “overcoming complex information problems, [thereby helping] willing governments to promote the inclusion of disadvantaged groups”, and has enthusiastically encouraged other governments to learn from the experience, and over 20 countries are reported to have expressed an interest in emulating Aadhaar.
It nevertheless remains controversial domestically. Critics have reportedly been harassed and surveilled for their opposition, and the scheme has been criticized for unnecessarily collecting biometric information, for severe shortcomings in legislative oversight, function creep, facilitating surveillance and other intrusions on the right to privacy, exacerbating cybersecurity issues, and creating barriers to accessing a range of social rights.
In 2018 India’s Supreme Court, in a 1448-page landmark ruling, upheld Aadhar’s constitutionality, albeit with some caveats. The court appeared to view the use of biometric identification technology in the context of providing welfare benefits as being legitimate, proportional and even inevitable. In a welfare state, Aadhaar’s aim of ensuring that benefits reach the intended beneficiary was “naturally a legitimate State aim”. In balancing the rights to social security and privacy, the court held that registering biometric data represented a ‘minimal’ inroad into privacy rights, and went so far as to characterize Aadhaar as “a vital tool of ensuring good governance in a social welfare state”. But the Supreme Court’s ruling has apparently not put an end to the controversy surrounding the scheme.
In 2019 Kenya required all of its citizens, including those living abroad, and all foreign nationals and refugees in the country, above the age of 6, to obtain a national ID in order to access government services, including welfare benefits. This involved providing biometric data including fingerprints, hand geometry, earlobe geometry, retina and iris patterns, voice waves and DNA in digital form. In response to a case claiming that this Huduma Namba program violated the rights to privacy, equality, non-discrimination and public participation, the High Court issued an interim order allowing the registration process to continue, but on a voluntary basis and on the basis that the disbursement of government services and benefits could not be made conditional on participation. Subsequently, registration has proceeded apace, some two-thirds of the population has been registered, and the Government is reportedly threatening to exclude unregistered individuals from access to benefits or the right to vote.
In South Africa, the South African Social Security Agency (SASSA), distributes non-contributory and means-tested social grants to around one-third of the population, including for example grants for child support, for pensioners, and for persons with disabilities. In 2012, SASSA contracted with the company Cash Paymaster Services (CPS), a subsidiary of Net1, to deliver the grants. CPS registered beneficiaries by collecting their biometric information (finger prints and, originally, voice recordings) and beneficiaries were issued MasterCard debit cards with biometric functionality and a linked bank account by Net1 and Grindrod Bank in association with SASSA. After much controversy surrounding the tender to CPS, the fees charged by CPS, deductions made to social grants on these accounts as well as privacy concerns surrounding the processing of card holder data, SASSA changed providers in 2018 by entering into a partnership with the South African Post Office (SAPO). SASSA and SAPO will provide new biometric cards. The change from CPS to SAPO has been complex and has led to questions about effective access to social grants by beneficiaries in South Africa.
Many other examples could be given of countries using or exploring digital identity systems, including Argentina, Bangladesh, Chile, Ireland, Jamaica, Malaysia, the Philippines, and the United States.
B. Eligibility assessment
Automated programs are increasingly used to assess eligibility in many countries. An especially instructive case was the automation of eligibility decisions in Ontario in 2014 through the Social Assistance Management System (SAMS) which relied on Cúram, a customizable off-the-shelf IBM software package, also used in welfare programs in Canada, the United States, Germany, Australia and New Zealand.
In 2015 the Ontario Auditor-General reported on 1,132 cases of errors with eligibility determinations and payment amounts under SAMS involving about $140 million. The total expenditure on SAMS by late 2015 was $290 million. The new system reportedly led caseworkers to resort to subterfuges to ensure that beneficiaries were fairly treated, made decisions very difficult to understand, and created significant additional work for staff.
C. Welfare benefit calculation and payments
The calculation and payment of benefits is increasingly done through digital technologies without the involvement of caseworkers and other human decision-makers. While such systems offer many potential advantages, the Special Rapporteur also received information about prominent examples of system errors or failures that generated major problems for large numbers of beneficiaries. These included the ‘Robodebt’ fiasco in Australia, the Real Time Information system in the United Kingdom, and the SAMS system in Canada.
Electronic payment cards or debit cards are also increasingly being issued to welfare recipients. Information provided to the Special Rapporteur in relation to such programs in New Zealand, Australia and South Africa reveal very similar problems. First, beneficiaries often face difficulties accessing and fully utilizing their right to social security. Second, when such cards are clearly recognizable as welfare-related, users have expressed feelings of disempowerment, embarrassment and shame, a problem exacerbated when the users come from communities long accustomed to exclusion. Third, electronic cards enable monitoring and surveillance of behavioral data by welfare authorities and private actors, thus raising important human rights concerns.
Fourth, the outsourcing of the issuance and administration of electronic cards to private companies has led to problems such as users being encouraged to pay for commercial financial products and the imposition of user fees. More generally the ethos surrounding such cards has often reflected stereotypes such as the financial untrustworthiness and irrationality of those living in poverty.
D. Fraud prevention and detection
Fraud and error in welfare systems can potentially involve very large sums of money and have long been a major concern for governments. It is thus unsurprising that many of the digital welfare systems that have been introduced have been designed with a particular emphasis on the capacity to match data from different sources in order to expose deception and irregularities on the part of welfare applicants. Nevertheless, evidence from country missions undertaken by the Special Rapporteur, along with other cases examined, suggests that the magnitude of these problems is frequently overstated and that there is sometimes a wholly disproportionate focus on this particular dimension of the complex welfare equation. Images of supposedly wholly undeserving individuals receiving large government welfare payments, such as Ronald Reagan’s ‘welfare queen’ trope, have long been used by conservative politicians to discredit the very concept of social protection. The risk is that the digital welfare state provides endless possibilities for taking surveillance and intrusion to new and deeply problematic heights.
E. Risk scoring and need classification
Risk calculation is inevitably at the heart of the design of welfare systems and digital technologies can achieve very high levels of sophistication in this regard. In addition to fraud detection and prevention, child protection has been a major focus in this area, as illustrated by examples from countries as diverse as the United States, New Zealand, the United Kingdom, and Denmark. Governments have also applied these techniques to determine whether unemployment assistance will be provided and at what level. A prominent such scheme in Poland was held unconstitutional, but an algorithm-based system in Austria continues to categorize unemployed jobseekers to determine the support they will receive from government jobcenters.
Many other areas of the welfare state will also be affected by new technologies used to score risks and classify needs. While such approaches offer many advantages, it is also important to take account of the problems that can arise. First there are many issues raised by determining an individual’s rights on the basis of predictions derived from the behavior of a general population group. Second, the functioning of the technologies and how they arrive at a certain score or classification is often secret, thus making it difficult to hold governments and private actors to account for potential rights violations. Third, risk-scoring and need categorization can reinforce or exacerbate existing inequalities and discrimination.
F. Communication between welfare authorities and beneficiaries
Communications that previously took place in person, by phone or by letter are increasingly being replaced by online applications and interactions. Various submissions to the Special Rapporteur cited problems with the Universal Credit system in the United Kingdom, including difficulties linked to a lack of internet access and/or digital skills, and the extent to which online portals can create confusion and obfuscate legal decisions, thereby undermining the right of claimants to understand and appeal decisions affecting their social rights. Similar issues were also raised in relation to other countries including Australia, and Greece.
Another problem is the likelihood that once the entire process of applying and maintaining benefits is moved online, the situation invites further digital innovation. In 2018 Sweden was forced to reverse a complex digital system used by the Employment Service to communicate with jobseekers because of problems that led to as many as 15% of the system’s decisions probably being incorrect.
Australia’s Targeted Compliance Framework (TCF) requires job seekers to interact with a digital dashboard to report mandatory activities and to check their compliance status. Failure to meet a ‘mutual obligation’ can automatically, without the involvement of a human decision-maker, lead to suspension of payments or the imposition of financial penalties. Submissions have highlighted problems due to a lack of internet access and digital literacy, to the rigidity of an automated system which fails to take real-life situations into account.
III. Making digital technologies work for social protection
Digital technologies, including artificial intelligence, have huge potential to promote the many benefits that are consistently cited by their proponents. They are already doing so for those who are economically secure and can afford to pay for the new services. They could also make an immense positive difference in improving the well-being of the less well-off members of society, but this will require deep changes in existing policies. The leading role in any such effort will have to be played by governments through appropriate fiscal policies and incentives, regulatory initiatives, and a genuine commitment to designing the digital welfare state not as a Trojan Horse for neoliberal hostility towards welfare and regulation but as a way to ensure a decent standard of living for everyone in society.
This report has sought to highlight problems that are specific to the ways in which the digital welfare state has been envisioned and implemented. But many of the changes required to avoid a digital dystopia will need to range more broadly. In addressing the General Assembly on 24 September 2019 the Prime Minister of the United Kingdom warned of the dangers of the digital age, singling out: (i) the risk of ‘round-the-clock surveillance’; (ii) the perils of algorithmic decision-making; (iii) the difficulty of appealing against computer-generated determinations; and (iv) the inability to plead extenuating circumstances when the decision-maker is an algorithm. He concluded rather ominously by suggesting that “[d]igital authoritarianism is … an emerging reality.”
His comments resonate strongly in the context of the digital welfare state, including in relation to the United Kingdom’s Universal Credit system. There is no magic recipe for avoiding the pitfalls of which he warned, but the following steps could help to make the digital welfare state a force for enhancing rather than undermining human rights.
A. Taking human rights seriously and regulating accordingly
The United Kingdom Prime Minister concluded his speech by warning that “[u]nless we ensure that new technology reflects” human rights, the Universal Declaration of Human Rights “will mean nothing … .” But the reality is that governments have certainly not regulated the tech industry as if human rights were at stake, and the technology sector remains virtually a human rights-free zone. The Big Tech companies and their governmental supporters have worked hard to keep it that way. Their approach can be summed up for present purposes in four propositions.
The first is that the ability to innovate requires freedom, especially from regulation. The Facebook founder’s early call for the industry to “move fast and break things” epitomizes the importance attached to minimizing legal and governmental constraints. Yet this argument leads inexorably to a handful of powerful executives replacing governments and legislators in determining the directions in which societies will move and the values and assumptions which will drive those developments. The accumulation of vast amounts of capital in the hands of very small elites and the rapid growth in extreme inequality have gone hand in hand with the ascendency of this approach so far.
The second proposition is that there are no universal values. In a recent book, the President of Microsoft asked rhetorically: “How can the world converge on a singular approach to ethics for computers when it cannot agree on philosophical issues for people?” Even non-discrimination standards are sometimes presented as being too vague and contested to be useful in regulating AI. But these arguments are self-serving and ill-informed. Governments worldwide have accepted universal human rights standards, including in the form of binding legal obligations. And over the past half century or more, these standards have been exhaustively developed and applied by courts, and a wide range of expert and community-based bodies. There remains plenty of room for philosophical disagreements, but there is no absence of agreement on core human values.
The third proposition is that governments are inherently slow and clumsy, and tend to respond to yesterday’s challenges rather than tomorrow’s. As the Republican minority leader of the United States House of Representatives recently argued, “the bureaucratic leviathan [does not have] what it takes to develop or enforce nimble responses to rapid change in the technology industry.” But while such claims might also be put forward by the proponents of unfettered discretion for the finance, aviation, defence, pharmaceutical, and other industries, it is solely in relation to Big Tech that governments have been prepared to abandon their regulatory responsibilities and acquiesce in a self-regulatory approach to such an extreme degree. There is no justification for such exceptionalism and no empirical evidence to support the claim that there is a fundamental incompatibility between innovation and regulation.
And the fourth proposition is that public accountability is unnecessary because the free market is the best regulator. Leaving aside the powerful arguments that Big Tech is deeply anti-competitive and thus immune to many currents of the free market, the great scandals of recent years that have led to the so-called ‘techlash’ provide compelling evidence that public accountability is indispensable.
In response to growing calls for effective governmental regulation, the industry has gone into high gear in producing, influencing and embracing ‘codes of ethics’ and other non-binding standards purporting to ‘regulate’ digital technologies and their developers. Most, but by no means all, of these codes contain a reference to human rights, but the substance of human rights law is invariably lacking. Instead the token reference to human rights serves only to enhance claims to legitimacy and universality. Meanwhile, the relevant discussions of ethics are based on almost entirely open-ended notions that are not necessarily grounded in legal or even philosophical arguments, and can be shaped to suit the needs of the industry. As a result, there are serious problems of conceptual incoherence, conflicts among norms are rarely acknowledged, meaningful input is rarely sought from stakeholders, and accountability mechanisms are absent. Even industry-employed ethicists acknowledge that “[i]f ethics is simply absorbed within the logics of market fundamentalism, meritocracy, and technological solutionism, it is unlikely that the tech sector will be able to offer a meaningful response to the desire for a more just and values-driven tech ecosystem.” Against this background, it is unsurprising that there are few public or scholarly discussions of the human rights implications of digital welfare states.
The human rights community has thus far done a very poor job of persuading industry, government, or seemingly society at large, of the fact that a technologically-driven future will be disastrous if it is not guided by respect for human rights that is in turn grounded in law.
B. Ensuring legality and transparency
One of the most surprising characteristics of too many important digital welfare state initiatives is a lack of attention to the importance of ensuring legality. Many examples have been drawn to the Special Rapporteur’s attention, including: the Australian Government’s online compliance intervention system which used automated data-matching as the basis to send out vast numbers of debt notices with very high error rates (known as Robodebt); allegedly unlawful information provided to claimants via the online Universal Credit portal in the United Kingdom; the contested legality of the Irish Public Services Card for some of the purposes for which it has been used; the SyRI (System Risk Indication) system in the Netherlands which initially lacked a legal basis and remains subject to court challenge; and the Aadhaar system in India which was originally implemented without a legal framework.
While the lack of a legal basis is deeply problematic per se, this gap also means that opportunities for legislative debate and for public inputs into shaping the relevant systems is also lacking. This has major potentially negative implications for transparency, design, legitimacy, and the likelihood of acceptance.
C. Promoting digital equality
Egalitarianism is a consistent theme of the technology industry, as exemplified by Facebook’s aim “to give people the power to build community and bring the world closer together”. But at the macro level Big Tech has been a driver of growing inequality and has facilitated the creation of a ‘vast digital underclass’.
For its part, the digital welfare state sometimes gives beneficiaries the option to go digital or continue using more traditional techniques. But in reality, policies such as ‘digital by default’ or ‘digital by choice’ are usually transformed into ‘digital only’ in practice. This in turn exacerbates or creates major disparities among different groups. A lack of digital literacy leads to an inability to use basic digital tools at all, let alone effectively and efficiently. Limited access, or no access to the internet, poses huge problems for a great many people. Additional barriers arise for individuals who have to pay high prices to obtain internet access, to travel long distances or absent themselves from work to do so, visit public facilities such as libraries in order to get access, or obtain assistance from staff or friends to navigate the systems. And while the well-off might have instant access to up-to-date, and easy to use computers and other hardware, as well as fast and efficient broadband speeds, the least well off are far more likely to be severely disadvantaged by out of date equipment and time-consuming and unreliable digital connections.
Submissions to the Special Rapporteur from a wide range of countries emphasized the salience of these different problems. Both in the Global North and the Global South, many individuals, and especially those living in poverty, do not have a reliable internet connection at home,  cannot afford such a connection, are not digitally skilled or confident, or are otherwise inhibited in communicating with authorities online. The various submissions emphasize how these problems impede the ability of would-be claimants to realize their human rights.
The United Kingdom provides an example of a wealthy country in which, even in 2019, 11.9 million people (22% of the population) do not have the ‘essential digital skills’ needed for day-to-day life. An additional 19% cannot perform fundamental tasks such as turning on a device or opening an app. In addition, 4.1 million adults (8%) are offline because of fears that the internet is an insecure environment, and proportionately almost half of those are from a low income household and almost half are under sixty years of age.
These problems are compounded by the fact that when digital technologies are introduced in welfare states, their distributive impact is often not a significant focus of governments. In addition, vulnerable individuals are not commonly involved in the development of IT systems and the IT professionals are often ill-equipped to anticipate the sort of problems that are likely to arise. Programs often assume, without justification, that individuals will have ready access to official documents and be able to upload them, that they will have a credit history or broader digital financial footprint, or even that their fingerprints will be readable, which is often not the case for those whose working lives have involved unremitting manual labour.
In terms of digital welfare policy, several conclusions emerge. First, there should always be a genuine non-digital option available. Second, programs that aim to digitize welfare arrangements should be accompanied by programs designed to promote and teach the needed digital skills and to ensure reasonable access to the necessary equipment as well as effective online access. Third, in order to reduce the harm caused by incorrect assumptions and mistaken design choices, digital welfare systems should be co-designed by their intended users and evaluated in a participatory manner.
D. Protecting economic and social rights in the digital welfare state
The processes of digitization and the increasing role played by automated decision-making through the use of algorithms and artificial intelligence have, in at least some respects, facilitated a move towards a bureaucratic process and away from one premised on the right to social security or the right to social protection. Rather than the ideal of the State being accountable to the citizen to ensure that the latter is able to enjoy an adequate standard of living, the burden of accountability has in many ways been reversed. To a greater degree than has often been the case in the past, today’s digital welfare state is often underpinned by the starting assumption that the individual is not a rights-holder but rather an applicant. In that capacity, a person must convince the decision-maker that they are ‘deserving’, that they satisfy the eligibility criteria, that they have fulfilled the often onerous obligations prescribed, and that they have no other means of subsistence. And much of this must be done electronically, regardless of the applicant’s skills in that domain.
The right to social security encompasses the right “to access and maintain benefits, whether in cash or in kind, without discrimination”. The imposition of technological requirements can make it impossible or very difficult for individuals to effectively access that right, thus making it effectively unaffordable.
The right to social protection is integrally linked to what the Human Rights Committee refers to as the right to life with dignity, which must be protected, where necessary, through “measures designed to ensure access without delay by individuals to essential goods and services such as food, water, shelter, health-care, electricity and sanitation, and other measures designed to promote and facilitate adequate general conditions … .” Various other rights are also implicated, including the right to an adequate standard of living, the right to mental health, and the right to be treated with dignity.
While social protection in general should be designed to protect these rights, the dignity dimension is at particular risk in the context of the digital welfare state. The potential risks arise in various contexts. First, the process for determining eligibility can easily be transformed into an electronic question and answer process that almost inevitably puts already vulnerable individuals at even greater disadvantage. Second, the way in which determinations are framed and communicated can be dehumanized and allow no room for meaningful questioning or clarification.
Third, the digital welfare state often seems to involve various forms of rigidity and the robotic application of rules. As a result, extenuating circumstances such as being late for an appointment because of urgent caring obligations, or being unable to understand a written communication whether because of a disability or a personal crisis, are often not taken into account in a predominantly digital context. Fourth, digital systems are often not designed to respond rapidly either to serious emergencies or to the daily challenges experienced by an older person whose entitlement has suddenly and inexplicably been electronically reduced or cancelled, or by a single parent unable to take a child to a local day care because the digital ID card will not function.
Fifth, the ways in which services are provided can easily have degrading connotations, such as unnecessarily exposing to a broader audience the fact that a person is reliant on benefits, or requiring extended waiting periods, or the navigation of lengthy queues. Sixth, the introduction of various new technologies that eliminate the human provider can enhance efficiency and provide other advantages, but might not necessarily be satisfactory for individuals who are in situations of particular vulnerability. New technologies often operate on the law of averages, in the interests of majorities, and on the basis of predicted outcomes or likelihoods.
Sixth, digital services risk eliminating, almost entirely, much of the human interaction and compassion that are likely to be indispensable components in providing at least some welfare recipients with the care and assistance they need. The assumption that there is always a technological fix for any problem is highly likely to be misplaced in various aspects of a humane and effective system of social protection.
E. Protecting civil and political rights in the digital welfare state
That the poor suffer from more intense levels of scrutiny, monitoring, surveillance, is hardly an original observation. In the 1960s, Charles Reich wrote that welfare recipients in the United States “have been subjected to many forms of procedure and control not imposed on other citizens. … [They] are all too easily regulated.” In 1975, Michel Foucault wrote about the ‘coercive technologies of behavior’ used in modern society to ‘discipline and punish’ the poorer classes.
By way of explaining why these lessons have not been learned in the digital welfare state, Shoshana Zuboff writes that the system of ‘surveillance capitalism’ that prevails today is unprecedented, which “has enabled it to elude systematic contest because it cannot be adequately grasped with our existing concepts.” This private surveillance is being reinforced by trends in government surveillance. Jack Balkin has described the ‘National Surveillance State’ as “a permanent feature of governance [that] will become as ubiquitous in time as the familiar devices of the regulatory and welfare states.”
Digital technologies are employed in the welfare state to surveil, target, harass and punish beneficiaries, especially the poorest and most vulnerable among them. Once again, many of the submissions received by the Special Rapporteur serve to illustrate and reinforce this point. They serve to highlight a number of human rights concerns. First, in the context of social security benefits and assistance, there is a real risk that beneficiaries are effectively forced to give up their right to privacy and data protection to receive their right to social security as well as other social rights.
A second concern is the blurring of the lines between public and private surveillance. Welfare state authorities increasingly rely, either actively or passively, on private corporations for the surveillance and targeting of beneficiaries. Private entities have different motives for their involvement in benefit and social assistance systems and this may lead to conflicts between the public interests these systems ought to serve and the private interests of corporations and their owners.
A third concern is the potential for deliberate targeting and harassment of the poor through new technologies in the welfare state. As one submission to the Special Rapporteur highlights, fraud in the welfare state is often the result of confusion, complexity and the inability to correct the resulting errors. But by deliberately using the power of new technologies to identify fraud or violations of ‘conditionalities’ imposed on beneficiaries, governments are likely to find inconsistencies that they can hold against claimants. It is relevant here that new technologies are enabling what Jack Balkin described as the ‘death of amnesia’: new abilities to collect information and store it digitally for an undefined period of time create a future in which a wealth of information can be held against someone indefinitely.
Additional concerns which warrant greater consideration than can be provided in the present report include (i) the human rights consequences of the move to predicting risk instead of the ex post enforcement of rules violations; (ii) the dangers of connecting government data siloes which is more readily contemplated in the welfare context than elsewhere in the field of digital governance; (iii) the psychological and societal cost of constant monitoring and surveillance; and (iv) the growing tendency of some governments to use the opportunities provided by the digital welfare state to try to alter social behaviours whether in the form of sexual activity or preferences, approaches to cohabitation, the use of alcohol or drugs, the decision to have children, or many other such goals.
F. Resisting the inevitability of a digital only future
Digital technologies in general, and especially those central to the digital welfare state, are often presented as being both unavoidable and irresistible. If a country wants to be seen to be at the technological cutting edge, if its government wants to have the most efficient, economical and flexible welfare system available, and if its citizenry wants all of the convenience that comes from not having to provide ID in order to undertake various transactions, then a transition to a digital welfare state must be pursued. But quite apart from the choices that citizens and governments might make if they were fully informed and adequately consulted, the reality is that such decisions are all too often taken in the absence of sophisticated cost-benefit analyses. And when such analyses are undertaken, they consist of financial balance sheets that ignore what might be termed the fiscally invisible intangibles that underpin human rights. Values such as dignity, choice, self-respect, autonomy, self-determination, privacy, and a range of other factors are all traded off without being factored into the overall equation, all but guaranteeing that insufficient steps will be taken to ensure their role in the new digital systems.
It is often assumed that at least some of these tradeoffs can be justified on the grounds that the bargain is just a matter between the individual and a particular government agency. But such an image is increasingly very far from the truth as cross-matching, data sharing, and cross-verification systematically enlarge the pools of data potentially available across the spectrum of government. To the extent that assurances are given that leakage from one silo to the next will not occur, such guarantees are largely illusory since a change of government or a real or imagined emergency situation is all that is required to trigger a partial or comprehensive breaking down of the partitions, quite apart from the risks of electronic data breaches due to hacking or normal system breakdowns. In addition, the assumption that the relationship is only that between government and citizen is also anachronistic. Corporate actors are now centrally involved in large parts of the welfare system, and when taken together with the ever-expanding reach of other forms of surveillance capitalism, intangible human rights values can be assumed to be worth as much as the shares of a bankrupt corporation.
The Special Rapporteur has learned of situations in which crucial decisions to go digital have been taken by government ministers without consultation, or even by departmental officials without any significant policy discussions taking place, on the grounds that the move is essentially an administrative matter, rather than involving a potentially game-changing approach to a large swathe of official policy. Sometimes there seems to be a presumption that even if the move to digital is not currently necessary, it surely will be one day and it is better to move in advance. Support for such pre-emptive moves may come from corporate interests, as well as from the security and counter-terrorism sectors, albeit for quite different reasons. Careful and transparent consideration should always be given to the questions of why, for whom, when, and how transitions to digital systems take place.
And even where detailed cost estimates are provided, this is an area in which accuracy seems difficult to achieve. Helen Margetts has observed that in the United Kingdom, for example, “[t]echnology and the public sector have rarely been happy bedfellows” and “every government technology project seems doomed to arrive late, underperform and come in over budget.” Another example is the Aadhaar system in India which is said to have lacked a proper cost-benefit analysis prior to implementation and in relation to which there has been great disagreement as to the post hoc assessment of costs and benefits.
 G. The role of the private sector
Two consistent themes of this report have been the reluctance of many governments to regulate the activities of technology companies, and the strong resistance of those companies to taking any systematic account of human rights considerations. The fact that this leads to many large technology corporations operating in an almost human rights free-zone is further exacerbated by the extent to which the private sector is taking a leading role in designing, constructing, and even operating significant parts of the digital welfare state.
Among well-known examples are the involvement of Net 1subsidiary Cash Paymaster Services (CPS), MasterCard and Grindrod Bank in the distribution of social grants linked to South Africa’s biometric identification system, the roles played by Indue and Visa in the Cashless Debit Card trials in Australia, and IBM’s involvement in Ontario’s Social Assistance Management System. Submissions to the Special Rapporteur have also drawn attention to the increasing role of the private sector in the German market for public administration software used for unemployment services, social and youth welfare; and outsourcing by United Kingdom local authorities to private companies in the area of social protection. In contrast, various submissions point to the deliberate choice by governments involved not to rely on private actors to play key roles in the welfare state.
The Special Rapporteur has addressed elsewhere the issues arising out of the privatization of public services more generally. But in relation to social protection services there is a deeply problematic lack of information about the precise role and responsibility of private actors in proposing, developing and operating digital technologies in welfare states around the world. This lack of transparency has a range of causes, from gaps in freedom of information laws, confidentiality clauses, and intellectual property protections, through a failure on the part of legislatures and executives to require transparency, to a general lack of investigation of these practices by oversight bodies and the media. The absence of information seriously impedes efforts to hold governments and private actors accountable.
H. Accountability mechanisms
Many of the programs used to promote the digital welfare state have been designed by the very same companies that are so deeply resistant to abiding by human rights standards. Moreover those companies and their affiliates are increasingly relied upon to design and implement key parts of the actual welfare programs. It is thus evident that the starting point for efforts to ensure human rights-compatible digital welfare states outcomes is to ensure through governmental regulation that technology companies are legally required to respect applicable international human rights standards.
IV. Conclusions
There is no shortage of analyses warning of the dangers for human rights of various manifestations of digital technology and especially artificial intelligence. But these studies focus overwhelmingly on the traditional civil and political rights such as the right to privacy, non-discrimination, fair trial rights, and the right to freedom of expression and information. With a handful of exceptions, none has adequately captured the full array of threats represented by the emergence of the digital welfare state. The vast majority of states spend very large amounts of money on different forms of social protection, or welfare, and the allure of digital systems that offer major cost savings along with personnel reductions, greater efficiency, and fraud reduction, not to mention the kudos associated with being at the technological cutting edge, is irresistible. There is little doubt that the future of welfare will be integrally linked to digitization and the application of AI.
But as humankind moves, perhaps inexorably, towards the digital welfare future it needs to alter course significantly and rapidly to avoid stumbling zombie-like into a digital welfare dystopia. Such a future would be one in which: unrestricted data matching is used to expose and punish the slightest irregularities in the record of welfare beneficiaries (while assiduously avoiding such measures in relation to the well-off); evermore refined surveillance options enable around the clock monitoring of beneficiaries; conditions are imposed on recipients that undermine individual autonomy and choice in relation to sexual and reproductive choices, and in relation to food, alcohol and drugs and much else; and highly punitive sanctions are able to be imposed on those who step out of line.
It will reasonably be objected that this report is unbalanced, or one-sided, because the dominant focus is on the risks rather than on the many advantages potentially flowing from the digital welfare state. The justification is simple. There are a great many cheerleaders extolling the benefits, but all too few counselling sober reflection on the downsides. Rather than seeking to summarize the analysis above, a number of additional observations are in order.
First, digital welfare state technologies are not the inevitable result of ‘scientific’ progress, but instead reflect political choices made by humans. Assuming that technology reflects pre-ordained or objectively rational and efficient outcomes risks abandoning human rights principles along with democratic decision-making.
Second, if the logic of the market is consistently permitted to prevail it inevitably disregards human rights considerations and imposes “externalities on society, for example when AI systems engage in bias and discrimination … and increasingly reduce human autonomy”.
Third, the values underpinning and shaping the new technologies are unavoidably skewed by the fact that there is “a diversity crisis in the AI sector across gender and race”. Those designing AI systems in general, as well as those focused on the welfare state are overwhelmingly white, male, well-off, and from the Global North. No matter how committed they might be to certain values, the assumptions and choices made in shaping the digital welfare state will reflect certain perspectives and life experiences. The way to counteract these biases and to ensure that human rights considerations are adequately taken into account is to ensure that the “practices underlying the creation, auditing, and maintenance of data” are subjected to very careful scrutiny.
Fourth, predictive analytics, algorithms and other forms of AI are highly likely to reproduce and exacerbate biases reflected in existing data and policies. In-built forms of discrimination can fatally undermine the right to social protection for key groups and individuals. There therefore needs to be a concerted effort to identify and counteract such biases in designing the digital welfare state. This in turn requires transparency, and broad-based inputs into policy-making processes. The public, and especially those directly affected by the welfare system, need to be able to understand and evaluate the policies that are buried deep within the algorithms.
Fifth, especially but not only in the Global North, the technology industry is heavily oriented towards designing and selling gadgets for the well-off such as driverless and flying cars and electronic personal assistants for multi-tasking businessmen [sic]. In the absence of fiscal incentives, government regulation, and political pressures, it will devote all too little attention to facilitating the creation of a welfare state that takes full account of the humanity and concerns of the less well-off in any society.
Sixth, to date astonishingly little attention has been paid to the ways in which new technologies might transform the welfare state for the better. Instead of obsessing about fraud, cost savings, sanctions, and market-driven definitions of efficiency, the starting point should be on how existing or even expanded welfare budgets could be transformed through technology to ensure a higher standard of living for the vulnerable and disadvantaged, to devise new ways of caring for those who have been left behind, and more effective techniques for addressing the needs of those who are struggling to enter or re-enter the labour market. That would be the real digital welfare state revolution.

Note, the footnotes which contain much of the meat in this report didnt get carried to this format. I highly recomment you hunt down the PDF version of this report which probably is the most failthful one. I'll try to find its URL. The Original that I have is a .doc file and I could export it as a PDF and upload it here if that is possible. My web site is down right now due to a denial of service attack. The UN's web site main URL is, I think, is uno.int The OCHR's web URL is https://www.ohchr.org/

up
1 user has voted.
zed2's picture

@zed2

Food security is the ultimate human right. Politics and politicians are to the starving, an unattainable unaffordable insulting and irrelevant luxury people cant even conceptualize if they dont have food, their children are dying, crying for milk and bodies are swollen from protein malnutrition.. History teaches us that starvation is the legacy of bad government. Famine is how a society treats those it deems unworthy of life. Increased climate variability and ENSO makes famines cyclic and unavoidable. The world has to set up means of providing people in drought afflicted areas food. The cult of the Invisible Hand that rules us views speculation on food during drought periods as desirable and legitimate. Hoever in a number of famines in recent years this dogma has proven to be hugely broken and wrong. Still it seems to have been accepted as truth during the huge Bengal Famine of 1945, when at least three million Indians died of starvation while citizens of the British Empire, arguably still the richest and most powerful nation on Earth. Here is the first, incomplete part of the UN special report by Philip Alston. Please try to read the full version, the footnotes are valuable and contain many links to must read works, (at least I consider them to be essential reading).

---------------------------------------

The report builds in part on country reports by the Special Rapporteur on visits to the United States in 2017 and the United Kingdom in 2018 which drew attention to the increasing use of digital technologies in social protection systems. In preparing the report, the Special Rapporteur consulted with representatives of various digital rights groups, leading scholars, and other stakeholders, first in a meeting hosted by the Digital Freedom Fund in Berlin in February 2019, and then at another sponsored by the Center for Information Technology Policy at Princeton University in April 2019. In addition, a formal call for contributions resulted in some 60 submissions from 34 countries, including 22 governments, as well as international and national civil society organizations, National Human Rights Institutions, academics, and individuals. While it is impossible to do justice to these rich and detailed submissions in such a necessarily brief report, the Special Rapporteur has made them electronically available and will continue analyzing them in the context of ongoing work on the digital welfare state.
II. Uses of digital technologies in the welfare state
From the many submissions received, and on the basis of various case studies addressed in the literature, it is possible to distinguish various ways, and different stages in the welfare context, in which digital innovation has been most prominently used.
A. Identity verification
Establishing every person’s legal identity, including through birth registration, by the year 2030 is the aim of Sustainable Development Goal (SDG) 16.9. A verifiable identity is essential for applying for benefits, establishing entitlement, receiving benefits, and appealing against denial. For the Government or other provider, it avoids duplication and fraud, it facilitates accurate targeting, and it enhances efficiency. Traditionally, paper and/or plastic documents have been used in forms such as birth certificates, identity cards and passports. These systems function reasonably well in most of the Global North, although 21 million adults in the United States do not have government-issued photo ID. In the Global South, 502 million people in sub-Saharan Africa and 357 million in South Asia lack official identification. In Liberia, for example, birth registration stands at only 5 per cent and national identity cards were not introduced until 2015.
In response, the World Bank, regional development organizations and bilateral donors have launched new programs to promote access to identity documents. In particular, the World Bank’s ‘ID4D’ (Identification for Development) campaign has focused heavily on promoting digital technologies as the key solution. This is explicitly stated in the ‘Principles on Identification for Sustainable Development’, which have been widely endorsed, including by MasterCard.
The Principles acknowledge that there are both pros and cons involved. On the positive side, it is claimed that digital technology can “create huge savings for citizens, governments, and businesses by reducing transaction costs, increasing efficiency, and driving innovation in service delivery, particularly to the poorest and most disadvantaged groups in society”. Digital identity systems can also “improve governance, boost financial inclusion, reduce gender inequalities by empowering women and girls, and increase access to health services and social safety nets for the poor”.
But in addition to this impressive and by now familiar sales pitch, the Principles, and similar documents, also recognize possible downsides ranging from political backlash to privacy and (cyber)security concerns. Solutions for dealing with those risks are often either technological or take the form of soft law norms. Thus a comparable United States Agency for International Development (USAID) document calls for “open source solutions” and developing good “practices for data privacy” to resolve the relevant problems. While the World Bank Principles refer to isolated human rights principles such as Article 7 of the Convention on the Rights of the Child, they rely primarily on the need to create an interoperable system using open standards, and protecting ‘privacy through system design’.
The world’s largest biometric identification system is Aadhaar in India, which is a 12-digit unique identifying number issued to Indian residents. It contains both demographic and biometric information, including an iris scan, a photograph and fingerprints. It is used to verify the identity of recipients of benefits and subsidies and is now mandatory to access those social rights. It was first introduced in 2009 and now covers more than 1.2 billion people, and has been enthusiastically endorsed by the international development community. The World Bank has praised it for “overcoming complex information problems, [thereby helping] willing governments to promote the inclusion of disadvantaged groups”, and has enthusiastically encouraged other governments to learn from the experience, and over 20 countries are reported to have expressed an interest in emulating Aadhaar.
It nevertheless remains controversial domestically. Critics have reportedly been harassed and surveilled for their opposition, and the scheme has been criticized for unnecessarily collecting biometric information, for severe shortcomings in legislative oversight, function creep, facilitating surveillance and other intrusions on the right to privacy, exacerbating cybersecurity issues, and creating barriers to accessing a range of social rights.
In 2018 India’s Supreme Court, in a 1448-page landmark ruling, upheld Aadhar’s constitutionality, albeit with some caveats. The court appeared to view the use of biometric identification technology in the context of providing welfare benefits as being legitimate, proportional and even inevitable. In a welfare state, Aadhaar’s aim of ensuring that benefits reach the intended beneficiary was “naturally a legitimate State aim”. In balancing the rights to social security and privacy, the court held that registering biometric data represented a ‘minimal’ inroad into privacy rights, and went so far as to characterize Aadhaar as “a vital tool of ensuring good governance in a social welfare state”. But the Supreme Court’s ruling has apparently not put an end to the controversy surrounding the scheme.
In 2019 Kenya required all of its citizens, including those living abroad, and all foreign nationals and refugees in the country, above the age of 6, to obtain a national ID in order to access government services, including welfare benefits. This involved providing biometric data including fingerprints, hand geometry, earlobe geometry, retina and iris patterns, voice waves and DNA in digital form. In response to a case claiming that this Huduma Namba program violated the rights to privacy, equality, non-discrimination and public participation, the High Court issued an interim order allowing the registration process to continue, but on a voluntary basis and on the basis that the disbursement of government services and benefits could not be made conditional on participation. Subsequently, registration has proceeded apace, some two-thirds of the population has been registered, and the Government is reportedly threatening to exclude unregistered individuals from access to benefits or the right to vote.
In South Africa, the South African Social Security Agency (SASSA), distributes non-contributory and means-tested social grants to around one-third of the population, including for example grants for child support, for pensioners, and for persons with disabilities. In 2012, SASSA contracted with the company Cash Paymaster Services (CPS), a subsidiary of Net1, to deliver the grants. CPS registered beneficiaries by collecting their biometric information (finger prints and, originally, voice recordings) and beneficiaries were issued MasterCard debit cards with biometric functionality and a linked bank account by Net1 and Grindrod Bank in association with SASSA. After much controversy surrounding the tender to CPS, the fees charged by CPS, deductions made to social grants on these accounts as well as privacy concerns surrounding the processing of card holder data, SASSA changed providers in 2018 by entering into a partnership with the South African Post Office (SAPO). SASSA and SAPO will provide new biometric cards. The change from CPS to SAPO has been complex and has led to questions about effective access to social grants by beneficiaries in South Africa.
Many other examples could be given of countries using or exploring digital identity systems, including Argentina, Bangladesh, Chile, Ireland, Jamaica, Malaysia, the Philippines, and the United States.
B. Eligibility assessment
Automated programs are increasingly used to assess eligibility in many countries. An especially instructive case was the automation of eligibility decisions in Ontario in 2014 through the Social Assistance Management System (SAMS) which relied on Cúram, a customizable off-the-shelf IBM software package, also used in welfare programs in Canada, the United States, Germany, Australia and New Zealand.
In 2015 the Ontario Auditor-General reported on 1,132 cases of errors with eligibility determinations and payment amounts under SAMS involving about $140 million. The total expenditure on SAMS by late 2015 was $290 million. The new system reportedly led caseworkers to resort to subterfuges to ensure that beneficiaries were fairly treated, made decisions very difficult to understand, and created significant additional work for staff.
C. Welfare benefit calculation and payments
The calculation and payment of benefits is increasingly done through digital technologies without the involvement of caseworkers and other human decision-makers. While such systems offer many potential advantages, the Special Rapporteur also received information about prominent examples of system errors or failures that generated major problems for large numbers of beneficiaries. These included the ‘Robodebt’ fiasco in Australia, the Real Time Information system in the United Kingdom, and the SAMS system in Canada.
Electronic payment cards or debit cards are also increasingly being issued to welfare recipients. Information provided to the Special Rapporteur in relation to such programs in New Zealand, Australia and South Africa reveal very similar problems. First, beneficiaries often face difficulties accessing and fully utilizing their right to social security. Second, when such cards are clearly recognizable as welfare-related, users have expressed feelings of disempowerment, embarrassment and shame, a problem exacerbated when the users come from communities long accustomed to exclusion. Third, electronic cards enable monitoring and surveillance of behavioral data by welfare authorities and private actors, thus raising important human rights concerns.
Fourth, the outsourcing of the issuance and administration of electronic cards to private companies has led to problems such as users being encouraged to pay for commercial financial products and the imposition of user fees. More generally the ethos surrounding such cards has often reflected stereotypes such as the financial untrustworthiness and irrationality of those living in poverty.
D. Fraud prevention and detection
Fraud and error in welfare systems can potentially involve very large sums of money and have long been a major concern for governments. It is thus unsurprising that many of the digital welfare systems that have been introduced have been designed with a particular emphasis on the capacity to match data from different sources in order to expose deception and irregularities on the part of welfare applicants. Nevertheless, evidence from country missions undertaken by the Special Rapporteur, along with other cases examined, suggests that the magnitude of these problems is frequently overstated and that there is sometimes a wholly disproportionate focus on this particular dimension of the complex welfare equation. Images of supposedly wholly undeserving individuals receiving large government welfare payments, such as Ronald Reagan’s ‘welfare queen’ trope, have long been used by conservative politicians to discredit the very concept of social protection. The risk is that the digital welfare state provides endless possibilities for taking surveillance and intrusion to new and deeply problematic heights.
E. Risk scoring and need classification
Risk calculation is inevitably at the heart of the design of welfare systems and digital technologies can achieve very high levels of sophistication in this regard. In addition to fraud detection and prevention, child protection has been a major focus in this area, as illustrated by examples from countries as diverse as the United States, New Zealand, the United Kingdom, and Denmark. Governments have also applied these techniques to determine whether unemployment assistance will be provided and at what level. A prominent such scheme in Poland was held unconstitutional, but an algorithm-based system in Austria continues to categorize unemployed jobseekers to determine the support they will receive from government jobcenters.
Many other areas of the welfare state will also be affected by new technologies used to score risks and classify needs. While such approaches offer many advantages, it is also important to take account of the problems that can arise. First there are many issues raised by determining an individual’s rights on the basis of predictions derived from the behavior of a general population group. Second, the functioning of the technologies and how they arrive at a certain score or classification is often secret, thus making it difficult to hold governments and private actors to account for potential rights violations. Third, risk-scoring and need categorization can reinforce or exacerbate existing inequalities and discrimination.
F. Communication between welfare authorities and beneficiaries
Communications that previously took place in person, by phone or by letter are increasingly being replaced by online applications and interactions. Various submissions to the Special Rapporteur cited problems with the Universal Credit system in the United Kingdom, including difficulties linked to a lack of internet access and/or digital skills, and the extent to which online portals can create confusion and obfuscate legal decisions, thereby undermining the right of claimants to understand and appeal decisions affecting their social rights. Similar issues were also raised in relation to other countries including Australia, and Greece.
Another problem is the likelihood that once the entire process of applying and maintaining benefits is moved online, the situation invites further digital innovation. In 2018 Sweden was forced to reverse a complex digital system used by the Employment Service to communicate with jobseekers because of problems that led to as many as 15% of the system’s decisions probably being incorrect.
Australia’s Targeted Compliance Framework (TCF) requires job seekers to interact with a digital dashboard to report mandatory activities and to check their compliance status. Failure to meet a ‘mutual obligation’ can automatically, without the involvement of a human decision-maker, lead to suspension of payments or the imposition of financial penalties. Submissions have highlighted problems due to a lack of internet access and digital literacy, to the rigidity of an automated system which fails to take real-life situations into account.
III. Making digital technologies work for social protection
Digital technologies, including artificial intelligence, have huge potential to promote the many benefits that are consistently cited by their proponents. They are already doing so for those who are economically secure and can afford to pay for the new services. They could also make an immense positive difference in improving the well-being of the less well-off members of society, but this will require deep changes in existing policies. The leading role in any such effort will have to be played by governments through appropriate fiscal policies and incentives, regulatory initiatives, and a genuine commitment to designing the digital welfare state not as a Trojan Horse for neoliberal hostility towards welfare and regulation but as a way to ensure a decent standard of living for everyone in society.
This report has sought to highlight problems that are specific to the ways in which the digital welfare state has been envisioned and implemented. But many of the changes required to avoid a digital dystopia will need to range more broadly. In addressing the General Assembly on 24 September 2019 the Prime Minister of the United Kingdom warned of the dangers of the digital age, singling out: (i) the risk of ‘round-the-clock surveillance’; (ii) the perils of algorithmic decision-making; (iii) the difficulty of appealing against computer-generated determinations; and (iv) the inability to plead extenuating circumstances when the decision-maker is an algorithm. He concluded rather ominously by suggesting that “[d]igital authoritarianism is … an emerging reality.”
His comments resonate strongly in the context of the digital welfare state, including in relation to the United Kingdom’s Universal Credit system. There is no magic recipe for avoiding the pitfalls of which he warned, but the following steps could help to make the digital welfare state a force for enhancing rather than undermining human rights.
A. Taking human rights seriously and regulating accordingly
The United Kingdom Prime Minister concluded his speech by warning that “[u]nless we ensure that new technology reflects” human rights, the Universal Declaration of Human Rights “will mean nothing … .” But the reality is that governments have certainly not regulated the tech industry as if human rights were at stake, and the technology sector remains virtually a human rights-free zone. The Big Tech companies and their governmental supporters have worked hard to keep it that way. Their approach can be summed up for present purposes in four propositions.
The first is that the ability to innovate requires freedom, especially from regulation. The Facebook founder’s early call for the industry to “move fast and break things” epitomizes the importance attached to minimizing legal and governmental constraints. Yet this argument leads inexorably to a handful of powerful executives replacing governments and legislators in determining the directions in which societies will move and the values and assumptions which will drive those developments. The accumulation of vast amounts of capital in the hands of very small elites and the rapid growth in extreme inequality have gone hand in hand with the ascendency of this approach so far.
The second proposition is that there are no universal values. In a recent book, the President of Microsoft asked rhetorically: “How can the world converge on a singular approach to ethics for computers when it cannot agree on philosophical issues for people?” Even non-discrimination standards are sometimes presented as being too vague and contested to be useful in regulating AI. But these arguments are self-serving and ill-informed. Governments worldwide have accepted universal human rights standards, including in the form of binding legal obligations. And over the past half century or more, these standards have been exhaustively developed and applied by courts, and a wide range of expert and community-based bodies. There remains plenty of room for philosophical disagreements, but there is no absence of agreement on core human values.
The third proposition is that governments are inherently slow and clumsy, and tend to respond to yesterday’s challenges rather than tomorrow’s. As the Republican minority leader of the United States House of Representatives recently argued, “the bureaucratic leviathan [does not have] what it takes to develop or enforce nimble responses to rapid change in the technology industry.” But while such claims might also be put forward by the proponents of unfettered discretion for the finance, aviation, defence, pharmaceutical, and other industries, it is solely in relation to Big Tech that governments have been prepared to abandon their regulatory responsibilities and acquiesce in a self-regulatory approach to such an extreme degree. There is no justification for such exceptionalism and no empirical evidence to support the claim that there is a fundamental incompatibility between innovation and regulation.
And the fourth proposition is that public accountability is unnecessary because the free market is the best regulator. Leaving aside the powerful arguments that Big Tech is deeply anti-competitive and thus immune to many currents of the free market, the great scandals of recent years that have led to the so-called ‘techlash’ provide compelling evidence that public accountability is indispensable.
In response to growing calls for effective governmental regulation, the industry has gone into high gear in producing, influencing and embracing ‘codes of ethics’ and other non-binding standards purporting to ‘regulate’ digital technologies and their developers. Most, but by no means all, of these codes contain a reference to human rights, but the substance of human rights law is invariably lacking. Instead the token reference to human rights serves only to enhance claims to legitimacy and universality. Meanwhile, the relevant discussions of ethics are based on almost entirely open-ended notions that are not necessarily grounded in legal or even philosophical arguments, and can be shaped to suit the needs of the industry. As a result, there are serious problems of conceptual incoherence, conflicts among norms are rarely acknowledged, meaningful input is rarely sought from stakeholders, and accountability mechanisms are absent. Even industry-employed ethicists acknowledge that “[i]f ethics is simply absorbed within the logics of market fundamentalism, meritocracy, and technological solutionism, it is unlikely that the tech sector will be able to offer a meaningful response to the desire for a more just and values-driven tech ecosystem.” Against this background, it is unsurprising that there are few public or scholarly discussions of the human rights implications of digital welfare states.
The human rights community has thus far done a very poor job of persuading industry, government, or seemingly society at large, of the fact that a technologically-driven future will be disastrous if it is not guided by respect for human rights that is in turn grounded in law.
B. Ensuring legality and transparency
One of the most surprising characteristics of too many important digital welfare state initiatives is a lack of attention to the importance of ensuring legality. Many examples have been drawn to the Special Rapporteur’s attention, including: the Australian Government’s online compliance intervention system which used automated data-matching as the basis to send out vast numbers of debt notices with very high error rates (known as Robodebt); allegedly unlawful information provided to claimants via the online Universal Credit portal in the United Kingdom; the contested legality of the Irish Public Services Card for some of the purposes for which it has been used; the SyRI (System Risk Indication) system in the Netherlands which initially lacked a legal basis and remains subject to court challenge; and the Aadhaar system in India which was originally implemented without a legal framework.
While the lack of a legal basis is deeply problematic per se, this gap also means that opportunities for legislative debate and for public inputs into shaping the relevant systems is also lacking. This has major potentially negative implications for transparency, design, legitimacy, and the likelihood of acceptance.
C. Promoting digital equality
Egalitarianism is a consistent theme of the technology industry, as exemplified by Facebook’s aim “to give people the power to build community and bring the world closer together”. But at the macro level Big Tech has been a driver of growing inequality and has facilitated the creation of a ‘vast digital underclass’.
For its part, the digital welfare state sometimes gives beneficiaries the option to go digital or continue using more traditional techniques. But in reality, policies such as ‘digital by default’ or ‘digital by choice’ are usually transformed into ‘digital only’ in practice. This in turn exacerbates or creates major disparities among different groups. A lack of digital literacy leads to an inability to use basic digital tools at all, let alone effectively and efficiently. Limited access, or no access to the internet, poses huge problems for a great many people. Additional barriers arise for individuals who have to pay high prices to obtain internet access, to travel long distances or absent themselves from work to do so, visit public facilities such as libraries in order to get access, or obtain assistance from staff or friends to navigate the systems. And while the well-off might have instant access to up-to-date, and easy to use computers and other hardware, as well as fast and efficient broadband speeds, the least well off are far more likely to be severely disadvantaged by out of date equipment and time-consuming and unreliable digital connections.
Submissions to the Special Rapporteur from a wide range of countries emphasized the salience of these different problems. Both in the Global North and the Global South, many individuals, and especially those living in poverty, do not have a reliable internet connection at home,  cannot afford such a connection, are not digitally skilled or confident, or are otherwise inhibited in communicating with authorities online. The various submissions emphasize how these problems impede the ability of would-be claimants to realize their human rights.
The United Kingdom provides an example of a wealthy country in which, even in 2019, 11.9 million people (22% of the population) do not have the ‘essential digital skills’ needed for day-to-day life. An additional 19% cannot perform fundamental tasks such as turning on a device or opening an app. In addition, 4.1 million adults (8%) are offline because of fears that the internet is an insecure environment, and proportionately almost half of those are from a low income household and almost half are under sixty years of age.
These problems are compounded by the fact that when digital technologies are introduced in welfare states, their distributive impact is often not a significant focus of governments. In addition, vulnerable individuals are not commonly involved in the development of IT systems and the IT professionals are often ill-equipped to anticipate the sort of problems that are likely to arise. Programs often assume, without justification, that individuals will have ready access to official documents and be able to upload them, that they will have a credit history or broader digital financial footprint, or even that their fingerprints will be readable, which is often not the case for those whose working lives have involved unremitting manual labour.
In terms of digital welfare policy, several conclusions emerge. First, there should always be a genuine non-digital option available. Second, programs that aim to digitize welfare arrangements should be accompanied by programs designed to promote and teach the needed digital skills and to ensure reasonable access to the necessary equipment as well as effective online access. Third, in order to reduce the harm caused by incorrect assumptions and mistaken design choices, digital welfare systems should be co-designed by their intended users and evaluated in a participatory manner.
D. Protecting economic and social rights in the digital welfare state
The processes of digitization and the increasing role played by automated decision-making through the use of algorithms and artificial intelligence have, in at least some respects, facilitated a move towards a bureaucratic process and away from one premised on the right to social security or the right to social protection. Rather than the ideal of the State being accountable to the citizen to ensure that the latter is able to enjoy an adequate standard of living, the burden of accountability has in many ways been reversed. To a greater degree than has often been the case in the past, today’s digital welfare state is often underpinned by the starting assumption that the individual is not a rights-holder but rather an applicant. In that capacity, a person must convince the decision-maker that they are ‘deserving’, that they satisfy the eligibility criteria, that they have fulfilled the often onerous obligations prescribed, and that they have no other means of subsistence. And much of this must be done electronically, regardless of the applicant’s skills in that domain.
The right to social security encompasses the right “to access and maintain benefits, whether in cash or in kind, without discrimination”. The imposition of technological requirements can make it impossible or very difficult for individuals to effectively access that right, thus making it effectively unaffordable.
The right to social protection is integrally linked to what the Human Rights Committee refers to as the right to life with dignity, which must be protected, where necessary, through “measures designed to ensure access without delay by individuals to essential goods and services such as food, water, shelter, health-care, electricity and sanitation, and other measures designed to promote and facilitate adequate general conditions … .” Various other rights are also implicated, including the right to an adequate standard of living, the right to mental health, and the right to be treated with dignity.
While social protection in general should be designed to protect these rights, the dignity dimension is at particular risk in the context of the digital welfare state. The potential risks arise in various contexts. First, the process for determining eligibility can easily be transformed into an electronic question and answer process that almost inevitably puts already vulnerable individuals at even greater disadvantage. Second, the way in which determinations are framed and communicated can be dehumanized and allow no room for meaningful questioning or clarification.
Third, the digital welfare state often seems to involve various forms of rigidity and the robotic application of rules. As a result, extenuating circumstances such as being late for an appointment because of urgent caring obligations, or being unable to understand a written communication whether because of a disability or a personal crisis, are often not taken into account in a predominantly digital context. Fourth, digital systems are often not designed to respond rapidly either to serious emergencies or to the daily challenges experienced by an older person whose entitlement has suddenly and inexplicably been electronically reduced or cancelled, or by a single parent unable to take a child to a local day care because the digital ID card will not function.
Fifth, the ways in which services are provided can easily have degrading connotations, such as unnecessarily exposing to a broader audience the fact that a person is reliant on benefits, or requiring extended waiting periods, or the navigation of lengthy queues. Sixth, the introduction of various new technologies that eliminate the human provider can enhance efficiency and provide other advantages, but might not necessarily be satisfactory for individuals who are in situations of particular vulnerability. New technologies often operate on the law of averages, in the interests of majorities, and on the basis of predicted outcomes or likelihoods.
Sixth, digital services risk eliminating, almost entirely, much of the human interaction and compassion that are likely to be indispensable components in providing at least some welfare recipients with the care and assistance they need. The assumption that there is always a technological fix for any problem is highly likely to be misplaced in various aspects of a humane and effective system of social protection.
E. Protecting civil and political rights in the digital welfare state
That the poor suffer from more intense levels of scrutiny, monitoring, surveillance, is hardly an original observation. In the 1960s, Charles Reich wrote that welfare recipients in the United States “have been subjected to many forms of procedure and control not imposed on other citizens. … [They] are all too easily regulated.” In 1975, Michel Foucault wrote about the ‘coercive technologies of behavior’ used in modern society to ‘discipline and punish’ the poorer classes.
By way of explaining why these lessons have not been learned in the digital welfare state, Shoshana Zuboff writes that the system of ‘surveillance capitalism’ that prevails today is unprecedented, which “has enabled it to elude systematic contest because it cannot be adequately grasped with our existing concepts.” This private surveillance is being reinforced by trends in government surveillance. Jack Balkin has described the ‘National Surveillance State’ as “a permanent feature of governance [that] will become as ubiquitous in time as the familiar devices of the regulatory and welfare states.”
Digital technologies are employed in the welfare state to surveil, target, harass and punish beneficiaries, especially the poorest and most vulnerable among them. Once again, many of the submissions received by the Special Rapporteur serve to illustrate and reinforce this point. They serve to highlight a number of human rights concerns. First, in the context of social security benefits and assistance, there is a real risk that beneficiaries are effectively forced to give up their right to privacy and data protection to receive their right to social security as well as other social rights.
A second concern is the blurring of the lines between public and private surveillance. Welfare state authorities increasingly rely, either actively or passively, on private corporations for the surveillance and targeting of beneficiaries. Private entities have different motives for their involvement in benefit and social assistance systems and this may lead to conflicts between the public interests these systems ought to serve and the private interests of corporations and their owners.
A third concern is the potential for deliberate targeting and harassment of the poor through new technologies in the welfare state. As one submission to the Special Rapporteur highlights, fraud in the welfare state is often the result of confusion, complexity and the inability to correct the resulting errors. But by deliberately using the power of new technologies to identify fraud or violations of ‘conditionalities’ imposed on beneficiaries, governments are likely to find inconsistencies that they can hold against claimants. It is relevant here that new technologies are enabling what Jack Balkin described as the ‘death of amnesia’: new abilities to collect information and store it digitally for an undefined period of time create a future in which a wealth of information can be held against someone indefinitely.
Additional concerns which warrant greater consideration than can be provided in the present report include (i) the human rights consequences of the move to predicting risk instead of the ex post enforcement of rules violations; (ii) the dangers of connecting government data siloes which is more readily contemplated in the welfare context than elsewhere in the field of digital governance; (iii) the psychological and societal cost of constant monitoring and surveillance; and (iv) the growing tendency of some governments to use the opportunities provided by the digital welfare state to try to alter social behaviours whether in the form of sexual activity or preferences, approaches to cohabitation, the use of alcohol or drugs, the decision to have children, or many other such goals.
F. Resisting the inevitability of a digital only future
Digital technologies in general, and especially those central to the digital welfare state, are often presented as being both unavoidable and irresistible. If a country wants to be seen to be at the technological cutting edge, if its government wants to have the most efficient, economical and flexible welfare system available, and if its citizenry wants all of the convenience that comes from not having to provide ID in order to undertake various transactions, then a transition to a digital welfare state must be pursued. But quite apart from the choices that citizens and governments might make if they were fully informed and adequately consulted, the reality is that such decisions are all too often taken in the absence of sophisticated cost-benefit analyses. And when such analyses are undertaken, they consist of financial balance sheets that ignore what might be termed the fiscally invisible intangibles that underpin human rights. Values such as dignity, choice, self-respect, autonomy, self-determination, privacy, and a range of other factors are all traded off without being factored into the overall equation, all but guaranteeing that insufficient steps will be taken to ensure their role in the new digital systems.
It is often assumed that at least some of these tradeoffs can be justified on the grounds that the bargain is just a matter between the individual and a particular government agency. But such an image is increasingly very far from the truth as cross-matching, data sharing, and cross-verification systematically enlarge the pools of data potentially available across the spectrum of government. To the extent that assurances are given that leakage from one silo to the next will not occur, such guarantees are largely illusory since a change of government or a real or imagined emergency situation is all that is required to trigger a partial or comprehensive breaking down of the partitions, quite apart from the risks of electronic data breaches due to hacking or normal system breakdowns. In addition, the assumption that the relationship is only that between government and citizen is also anachronistic. Corporate actors are now centrally involved in large parts of the welfare system, and when taken together with the ever-expanding reach of other forms of surveillance capitalism, intangible human rights values can be assumed to be worth as much as the shares of a bankrupt corporation.
The Special Rapporteur has learned of situations in which crucial decisions to go digital have been taken by government ministers without consultation, or even by departmental officials without any significant policy discussions taking place, on the grounds that the move is essentially an administrative matter, rather than involving a potentially game-changing approach to a large swathe of official policy. Sometimes there seems to be a presumption that even if the move to digital is not currently necessary, it surely will be one day and it is better to move in advance. Support for such pre-emptive moves may come from corporate interests, as well as from the security and counter-terrorism sectors, albeit for quite different reasons. Careful and transparent consideration should always be given to the questions of why, for whom, when, and how transitions to digital systems take place.
And even where detailed cost estimates are provided, this is an area in which accuracy seems difficult to achieve. Helen Margetts has observed that in the United Kingdom, for example, “[t]echnology and the public sector have rarely been happy bedfellows” and “every government technology project seems doomed to arrive late, underperform and come in over budget.” Another example is the Aadhaar system in India which is said to have lacked a proper cost-benefit analysis prior to implementation and in relation to which there has been great disagreement as to the post hoc assessment of costs and benefits.
 G. The role of the private sector
Two consistent themes of this report have been the reluctance of many governments to regulate the activities of technology companies, and the strong resistance of those companies to taking any systematic account of human rights considerations. The fact that this leads to many large technology corporations operating in an almost human rights free-zone is further exacerbated by the extent to which the private sector is taking a leading role in designing, constructing, and even operating significant parts of the digital welfare state.
Among well-known examples are the involvement of Net 1subsidiary Cash Paymaster Services (CPS), MasterCard and Grindrod Bank in the distribution of social grants linked to South Africa’s biometric identification system, the roles played by Indue and Visa in the Cashless Debit Card trials in Australia, and IBM’s involvement in Ontario’s Social Assistance Management System. Submissions to the Special Rapporteur have also drawn attention to the increasing role of the private sector in the German market for public administration software used for unemployment services, social and youth welfare; and outsourcing by United Kingdom local authorities to private companies in the area of social protection. In contrast, various submissions point to the deliberate choice by governments involved not to rely on private actors to play key roles in the welfare state.
The Special Rapporteur has addressed elsewhere the issues arising out of the privatization of public services more generally. But in relation to social protection services there is a deeply problematic lack of information about the precise role and responsibility of private actors in proposing, developing and operating digital technologies in welfare states around the world. This lack of transparency has a range of causes, from gaps in freedom of information laws, confidentiality clauses, and intellectual property protections, through a failure on the part of legislatures and executives to require transparency, to a general lack of investigation of these practices by oversight bodies and the media. The absence of information seriously impedes efforts to hold governments and private actors accountable.
H. Accountability mechanisms
Many of the programs used to promote the digital welfare state have been designed by the very same companies that are so deeply resistant to abiding by human rights standards. Moreover those companies and their affiliates are increasingly relied upon to design and implement key parts of the actual welfare programs. It is thus evident that the starting point for efforts to ensure human rights-compatible digital welfare states outcomes is to ensure through governmental regulation that technology companies are legally required to respect applicable international human rights standards.
IV. Conclusions
There is no shortage of analyses warning of the dangers for human rights of various manifestations of digital technology and especially artificial intelligence. But these studies focus overwhelmingly on the traditional civil and political rights such as the right to privacy, non-discrimination, fair trial rights, and the right to freedom of expression and information. With a handful of exceptions, none has adequately captured the full array of threats represented by the emergence of the digital welfare state. The vast majority of states spend very large amounts of money on different forms of social protection, or welfare, and the allure of digital systems that offer major cost savings along with personnel reductions, greater efficiency, and fraud reduction, not to mention the kudos associated with being at the technological cutting edge, is irresistible. There is little doubt that the future of welfare will be integrally linked to digitization and the application of AI.
But as humankind moves, perhaps inexorably, towards the digital welfare future it needs to alter course significantly and rapidly to avoid stumbling zombie-like into a digital welfare dystopia. Such a future would be one in which: unrestricted data matching is used to expose and punish the slightest irregularities in the record of welfare beneficiaries (while assiduously avoiding such measures in relation to the well-off); evermore refined surveillance options enable around the clock monitoring of beneficiaries; conditions are imposed on recipients that undermine individual autonomy and choice in relation to sexual and reproductive choices, and in relation to food, alcohol and drugs and much else; and highly punitive sanctions are able to be imposed on those who step out of line.
It will reasonably be objected that this report is unbalanced, or one-sided, because the dominant focus is on the risks rather than on the many advantages potentially flowing from the digital welfare state. The justification is simple. There are a great many cheerleaders extolling the benefits, but all too few counselling sober reflection on the downsides. Rather than seeking to summarize the analysis above, a number of additional observations are in order.
First, digital welfare state technologies are not the inevitable result of ‘scientific’ progress, but instead reflect political choices made by humans. Assuming that technology reflects pre-ordained or objectively rational and efficient outcomes risks abandoning human rights principles along with democratic decision-making.
Second, if the logic of the market is consistently permitted to prevail it inevitably disregards human rights considerations and imposes “externalities on society, for example when AI systems engage in bias and discrimination … and increasingly reduce human autonomy”.
Third, the values underpinning and shaping the new technologies are unavoidably skewed by the fact that there is “a diversity crisis in the AI sector across gender and race”. Those designing AI systems in general, as well as those focused on the welfare state are overwhelmingly white, male, well-off, and from the Global North. No matter how committed they might be to certain values, the assumptions and choices made in shaping the digital welfare state will reflect certain perspectives and life experiences. The way to counteract these biases and to ensure that human rights considerations are adequately taken into account is to ensure that the “practices underlying the creation, auditing, and maintenance of data” are subjected to very careful scrutiny.
Fourth, predictive analytics, algorithms and other forms of AI are highly likely to reproduce and exacerbate biases reflected in existing data and policies. In-built forms of discrimination can fatally undermine the right to social protection for key groups and individuals. There therefore needs to be a concerted effort to identify and counteract such biases in designing the digital welfare state. This in turn requires transparency, and broad-based inputs into policy-making processes. The public, and especially those directly affected by the welfare system, need to be able to understand and evaluate the policies that are buried deep within the algorithms.
Fifth, especially but not only in the Global North, the technology industry is heavily oriented towards designing and selling gadgets for the well-off such as driverless and flying cars and electronic personal assistants for multi-tasking businessmen [sic]. In the absence of fiscal incentives, government regulation, and political pressures, it will devote all too little attention to facilitating the creation of a welfare state that takes full account of the humanity and concerns of the less well-off in any society.
Sixth, to date astonishingly little attention has been paid to the ways in which new technologies might transform the welfare state for the better. Instead of obsessing about fraud, cost savings, sanctions, and market-driven definitions of efficiency, the starting point should be on how existing or even expanded welfare budgets could be transformed through technology to ensure a higher standard of living for the vulnerable and disadvantaged, to devise new ways of caring for those who have been left behind, and more effective techniques for addressing the needs of those who are struggling to enter or re-enter the labour market. That would be the real digital welfare state revolution.

Note, the footnotes which contain much of the meat in this report didnt get carried to this format. I highly recomment you hunt down the PDF version of this report which probably is the most failthful one. I'll try to find its URL. The Original that I have is a .doc file and I could export it as a PDF and upload it here if that is possible. My web site is down right now due to a denial of service attack. The UN's web site main URL is, I think, is uno.int The OCHR's web URL is https://www.ohchr.org/

up
1 user has voted.
QMS's picture

@zed2

There are a lot of words there. Wink

up
3 users have voted.

@zed2 hi, sorry to hear your site is having troubles. It would help me a lot if you could trim your C99 comments to the standard Fair Use three paragraphs plus your opinions, and include the link to the source at the top. Instead of pasting three giant walls of text that makes me go blind, I can't read it. This isn't the proper format to be hosting all that text in one comment or three, in my opinion. It's more essay-like. Not the content, the format.

Thanks for all the information, even though it is hard for me to digest I still try. You remind me of another commenter on The Other Site from a long time ago, they really had great sources too. I am down with organic community so don't stop posting, that's not what I mean. I hope you get what I mean, it is from the heart but also from the poor eyes. trim baby trim

Peace and Love
eternal september

up
6 users have voted.
zed2's picture

@eyo @eyo @eyo
much better. I have multiple eye diseases, officially and now they are in retreat. I managed to fix some of it. I'm not kidding!

Maybe I have some experience that might help you? Anyway, here is why I started pasting things where that is clearly okay in (you'll notice I never do that to anything encumbered by a copyright) In the case of UN Special Reports it is all open access. I the case of other content, its usually open access or in some cases, that the author personally told me in writing that it was okay to share. But mostly you'll notice that virtually nothing I post links to is behind a paywall or encumbered. I am poor myself, and everything I read is free, pretty much. Others dont charge me to read it and its free to read. There is an NGO, creative commons, that a friend sits on the board of that brings the concept of open licensing to online content. A lot of content carries a Creative Commons license, do you know what that is? People like the DNC - as they work for the corporate takeover of our planet are foolishly trying to eliminate open licensing, trying to make up every possible excuse they can to censor voices because they know the world is wising up to their thefts. So, I paste it because I can tell that people rarely if ever follow my links because they clearly dont read these important things that they explain. If people discussed them and or integrated the facts explained in them as they used to in the past, Before 2013 Before the government started engaging in sockpuppetry? They are stealing the taxpayers money in order to try to influence us in political issues that its up to VOTERS not politicos to decide. They know we the voterss reject their BS, so now they are trying to nullify the peoples wishes and hde any facts that arent pablum like they emit.

They hate the concept that some people may want to share their ideas and hard work's products with the world (on the condition that others share their improvements too, sometimes) Tools that they made themselves and not charge people for using it making it unavailable to the poor. If they were real Democrats they would be telling us all how much free high quality content and software was being given away free to all of us TO USE. (This could revolutionize our schools) But they don't, they are trying to HIDE THIS and STOP - yes- STOP it!

Well they can go to hell. If it wasn't for FOSS, none of the economic growth of the 30 years would even exist! We would not have Google or Facebook or Apple or Netflix - They all happened and could not have happened without very high quality free software, being given away for free. Smart people built those companies on top of these free tools and they owe everything to them. Without FOSS their founders will tell you, they would not exist. Really. Ask them.

The government wasn't always this captured by idiots. I and many others actually learned about open source from them. Thats where it originated. What the hell has happened there?

up
1 user has voted.
zed2's picture

They are consolidating their power right now. We should be aware of that, we're in serious denial!

We should bone up on the international law that regulates governments ability to provide services to poor people (or other people) Which may be viewed by business as a taking from them of an entitlement to do business. It definitely is now.

I just found a video which I have not seen before by a German lawyer Markus Krajewski who is one of the experts on the clash between public services and the new neoliberal ruling ideology.

If you insidt on denying that everything has been changed and that as some politicians pretend, that its still 1980 - in other words, before any of the FTAs, youre doomed to never understand this new world. If you snooze, we all lose, Big.

These FTAs changed everything. The rich took back the world and nullified democracy. Is this legitimate? They insist it is. They tell us it is and refuse to discuss it any further.

up
4 users have voted.
The Liberal Moonbat's picture

@zed2
Consent.
Of.
The.
Governed.

They "refuse to discuss it any further", we have every right to kill them. Simple.

This is 1776 all over again - funny how the Founding Fathers should get "cancelled" for moral impurity (of which they themselves were well aware, TYVM) just as the scenario starts to look EXACTLY like the one they faced, isn't it? I recall reading a bunch of good articles about parallels to that period in the Occupy era before the NewTruth tsunami.

up
1 user has voted.

In the Land of the Blind, the One-Eyed Man is declared mentally ill for describing colors.

Yes Virginia, there is a Global Banking Conspiracy!

zed2's picture

For well to do people only. So they try to keep powerless people poor. On the edge of starvation, often.

People that lose political power usually lose all their rights and property in this world. History tells us that. Housing, jobs, lands, possessions of all kinds, are taken, and sold (usually) or given to others.

Isn't that whats happening now?

up
4 users have voted.
zed2's picture

The fitting end of the Nazi movement, as the citizens of Munich welcomed the end of the "Thousand year Reich" and the beginning of a new era of prosperity and peace for Germany.

Although at the beginning food was hard to find and expensive, the efforts of many enterprising Germans such as Theo and Karl Albrecht, brought good food to Germans in the almost totally devastated country quickly.

up
4 users have voted.
gulfgal98's picture

as to how much the American people are willing to tolerate from the oligarchs. Every straw appears to be the last straw until the oligarchy finds another one to pile up on the backs of the people, especially the poor.

up
1 user has voted.

Do I hear the sound of guillotines being constructed?

“Those who make peaceful revolution impossible will make violent revolution inevitable." ~ President John F. Kennedy