NEWSLETTER SIGNUP

Do Something

Make your voice heard

Find out who your state and federal representatives are and reach out. You can also find out who represents you on the City Council and contact them. Technology is changing rapidly, and the legislative and judicial systems are slow to keep up. Tell your representatives and those making policy that we must keep human beings at the helm of services and decision-making.

Connect WITH OUR SOCIAL ACTION TEAM



Want more of The Citizen?

Sign up for our newsletter

For a weekly dose of ideas, solutions and practical action steps, sign up for our newsletter:

* indicates required



/

( mm / dd )


And follow us on Facebook, Twitter & Instagram.

Be a better Philadelphia Citizen

Here's how

One of the founding tenets of The Philadelphia Citizen is to get people the resources they need to become better, more engaged citizens of their city.

We hope to do that in our Good Citizenship Toolkit, which includes a host of ways to get involved in Philadelphia — whether you want to contact your City Councilmember about safety incorporating AI into human services, get those experiencing homelessness the goods they need, or simply go out to dinner somewhere where you know your money is going toward a greater good.

Find an issue that’s important to you in the list below, and get started on your journey of A-plus citizenship.

Vote and strengthen democracy

Stand up for marginalized communities

Create a cleaner, greener Philadelphia

Help our local youth and schools succeed

Support local businesses

LISTEN

To this story in CitizenCast

Welcome to the enhanced audio edition of Trooper Sanders’s story


And go here for more audio articles from CitizenCast

Guest Commentary: Humans Must Control Human-Serving AI

Benefits Data Trust’s CEO — an AI advisor to the president — on what we must remember when using machines to deliver services to humans

Guest Commentary: Humans Must Control Human-Serving AI

Benefits Data Trust’s CEO — an AI advisor to the president — on what we must remember when using machines to deliver services to humans

For leaders of human services agencies delivering food, health care, and other critical assistance to people and families in need, the fizzy talk about artificial intelligence (AI) may seem enticing. Dedicated public servants on the front lines of fighting poverty, illness, and exclusion made it through the unprecedented demands of the Covid-19 pandemic only to find themselves adjusting to a new normal — serving their communities while severely understaffed, overstretched, and reliant upon systems and practices from a bygone era.

The purveyors of AI say their wares will bring an unprecedented boost to productivity, knowledge, and wealth creation. The truth is that AI brings many benefits, but it is wise for governments and nonprofits to proceed with caution. The promise of AI is tempered by well-known risks such as exacerbating bias and discrimination, contributing to a toxic workplace, and menacing privacy. These failings should be of particular concern to the leaders of human services agencies.

Indeed, rogue code can upend an agency’s mission to improve and save lives. An ongoing scandal in Australia tells a cautionary tale.

AI fails Australia

In 2015, the Australian government used AI to turbocharge efforts to track and recover suspected fraudulent social security payments. The cabinet minister in charge of social services at the time said he would be a “strong welfare cop on the beat” bringing cheats to heel. The government’s “Robodebt” effort tapped an automated decision-making system, but the system was not fit for this purpose.

In an exhaustive report, the Royal Commission that investigated the Robodebt scheme found gross inaccuracies in the system’s automated flags, unreasonable demands placed on the accused and, most tragically, lives lost. “Robodebt was a crude and cruel mechanism, neither fair nor legal, and it made many people feel like they were criminals,” the Commission concluded. “In essence, people were traumatized on the off chance they might owe money. It was a costly failure of public administration, in both human and economic terms.”

AI does not think. That’s an exclusively human power … It can augment human expertise but not replace it.

There is a growing body of work, including the U.S. federal government’s Blueprint for an AI Bill of Rights and AI Risk Management Framework, as well as a range of tools on trustworthy AI produced by civil society groups that can help human services agencies chart a responsible AI course.

But the key lesson from Australia is that leadership matters. Leaders can start by ensuring AI decisions are part of agency-wide strategy and policy instead of being left to data and technology experts or the back office. Then they must set the right tone. Political posturing about being the high sheriff of waste, fraud, and abuse signals technical and procurement officials to shop for AI products that are little more than algorithmic nightsticks. To be sure, program integrity should be a top priority, but true program integrity must both keep the ineligible from drawing down assistance while also ensuring those who are eligible can get help efficiently, with dignity, and free from unwarranted suspicion.

The human-in-the-loop principle

Human services leaders can look to the military for a key principle on managing AI: keeping the human-in-the-loop. Put simply, advances in AI and military capabilities are giving weapons the ability to select and engage targets — i.e., kill — without human intervention. This technical reality can be checked by policies that require humans stay at the controls of lethal force. Whether human control can keep pace with autonomous capabilities is an open question, but establishing the human-in-the-loop principle is a critical, and potentially life-saving, first step.

A human-in-the-loop principle for human services would cover two affected parties: front-line workers and the people served by programs. What’s key to remember is that AI does not think. That’s an exclusively human power. Where low-risk, rote tasks consume too much human bandwidth, AI can be an important efficiency tool, and should cautiously be given more room to operate. But when AI affects the rights, safety and well-being of people and society, such as determining eligibility for public benefits or adjudicating disputes, AI should be kept on a short, tightly controlled leash. It can augment human expertise but not replace it.

In Australia’s case, before Robodebt, payments flagged as fraudulent were reviewed manually, with compliance officers contacting recipients to see if discrepancies could be resolved using the rules, human judgment, and common sense. Robodebt shoved skilled workers to the sidelines with tragic consequences.

When designing AI strategies, human services leaders should bring experts and frontline workers into the mix early, tailoring requirements for AI products to the lived experience of workers, and they should ensure workers remain in the loop throughout the lifecycle of its procurement and use.

For example, as part of our work closing the nation’s more than $80 billion benefits access gap, Benefits Data Trust worked closely with financial aid experts, college advisors, and students to guide our use of natural language processing, a branch of AI, to power the chatbot Wyatt that advises rising college students through the federal financial aid application.

The people seeking and receiving services should have a seat at the table as well. This is best done by bringing human-centered design and the voice and interests of the client into the heart of AI planning. Centering the best interests of people is a useful screen for AI use cases and purchasing requirements, as well as evaluating the technology’s impact on an agency’s work.

The Australian review found no evidence that the creators of Robodebt attempted to consider how the vulnerable and disadvantaged would experience a wholly automated process. Human-centered design could have avoided the catastrophe by not only engaging in public consultation but changing the way the agency worked as well. As Civilla, a nonprofit with expertise in human-centered design, advises: “successful human-centered design in public institutions requires us to manage change across complex organizations — planning for ongoing evaluations, developing internal capacity, and guiding culture change at scale.”

Tapping the wisdom of both workers and residents alike can dramatically increase the chances of getting the best out of AI and mitigating the worst. Trustworthy AI can not only introduce much-needed technical muscle to important government programs, thoughtful strategies guiding its use can also drive change that brings more efficiency and dignity to human services and delivers more value to the taxpayer. Making it happen is all a question of leadership.


Trooper Sanders is CEO of Benefits Data Trust, a Philadelphia-based national nonprofit that uses data, technology, policy change, and direct service to help people tap the more than $80 billion in unclaimed benefits that support critical needs such as food security and healthcare. He sits on the National AI Advisory Committee. This piece originally ran in Tech Policy Press.

MORE TECHNOLOGY COVERAGE FROM THE CITIZEN

Photo by Possessed Photography on Unsplash

The Philadelphia Citizen will only publish thoughtful, civil comments. If your post is offensive, not only will we not publish it, we'll laugh at you while hitting delete.

Be a Citizen Editor

Suggest a Story

Advertising Terms

We do not accept political ads, issue advocacy ads, ads containing expletives, ads featuring photos of children without documented right of use, ads paid for by PACs, and other content deemed to be partisan or misaligned with our mission. The Philadelphia Citizen is a 501(c)(3) nonprofit, nonpartisan organization and all affiliate content will be nonpartisan in nature. Advertisements are approved fully at The Citizen's discretion. Advertisements and sponsorships have different tax-deductible eligibility. For questions or clarification on these conditions, please contact Director of Sales & Philanthropy Kristin Long at [email protected] or call (609)-602-0145.