inkwell.vue.505 : Virginia Eubanks, Automating Inequality
permalink #0 of 43: Jon Lebkowsky (jonl) Tue 4 Sep 18 14:39
    
Inkwell welcomes Virginia Eubanks, author of  Automating Inequality:
How High-Tech Tools Profile, Police, and Punish the Poor. Virginia
is an Associate Professor of Political Science at the University at
Albany, SUNY. She is also the author of Digital Dead End: Fighting
for Social Justice in the Information Age; and co-editor, with
Alethia Jones, of Ain’t Gonna Let Nobody Turn Me Around: Forty Years
of Movement Building with Barbara Smith. Her writing about
technology and social justice has appeared in The American Prospect,
The Nation, Harper’s and Wired. For two decades, Eubanks has worked
in community technology and economic justice movements. Today, she
is a founding member of the Our Data Bodies Project and a Fellow at
New America. She lives in Troy, NY.

Leading the interview is Ari Davidow, who has a background in
community planning and software/web development. Ari has been
working with online community for many years, including a long
history as community host here on the WELL. He has also been a host
at other online communities, mailing lists, blogs and websites.
  
inkwell.vue.505 : Virginia Eubanks, Automating Inequality
permalink #1 of 43: Ari Davidow (ari) Tue 4 Sep 18 20:26
    
Virginia's most recent book is "Automating Inequality," which was 
recommended a few months ago here on the WELL. In the book, she covers 
three primary case studies showing how we are using big data and AI in 
ways that, as the book title "automate inequality."

The book hit me harder than I expected. Since the 2016 election, I have 
been focused on ways that technology can help us organize more 
effectively. I have been slow to realize that both inadvertently and 
otherwise, we have long been making the lives of the underserved more 
difficult, while also letting ourselves be seduced by the apparent 
"all-knowingness" of big data and machine learning to do even worse.

Perhaps the best way to start things off, though, is to let Eubanks talk 
about how she got involved in doing this work--both her work with the "Our 
Data Bodies" project (http://www.odbproject.org), and how she came to 
write this book.
  
inkwell.vue.505 : Virginia Eubanks, Automating Inequality
permalink #2 of 43: Ari Davidow (ari) Thu 6 Sep 18 08:20
    
For those who haven't heard of the book, here are some resources
that cover the book and the issues surrounding it:

NPR story, Feb 19, 2018, "'Automating Inequality': Algorithms In
Public Services Often Fail The Most Vulnerable":
<https://n.pr/2GsR6Vy>

Discussion w/Virgina Eubanks on PBS, Jan 16, 2018:
<https://to.pbs.org/2CoiFSL>

Discussion between Virginia Eubanks and Kathryn Edin, author of "$2
a Day," on CSPAN, Mar 29, 2018: <https://cs.pn/2wNTeUF>
  
inkwell.vue.505 : Virginia Eubanks, Automating Inequality
permalink #3 of 43: Julie Rehmeyer (jrehmeyer) Thu 6 Sep 18 10:47
    
I'm very excited to see this discussion, because I've been
interested in this book for a while but haven't gotten a chance to
read it.

I was just watching the PBS show, and I was really struck by this
comment, at 23:00. Eubanks is asked what we need to do to solve this
problem, and she says that the first thing is "get our souls right
around poverty in the United States, because as long as we believe
that folks are poor because they've made bad choices, as long as we
believe that poverty is an aberration and not the majority
experience int he United States, we're going to produce these
systems that -- no matter how hard we try to be fair -- reproduce
politics as usual."

As most folks here know, I've done a lot of work around poorly
understood illnesses, and I'm struck by the same dynamics in the
conversation there. Sick people, and most especially people who have
the nerve to get sick with illnesses that are not 100% accepted, are
seen as victims of their own bad choices. That's the gist of the
argument that these illnesses are "all in our heads." And our
experience is othered -- if we're sick because we're actually crazy,
then normal people need not worry that they'll ever get sick like
us. In a similar way, seeing poverty as an aberration makes those
who aren't poor feel safe -- just be a normal person and you'll be
fine.

Sorry to come in early with a comment that's a bit to the side of
her main argument, but her way of formulating the problem was useful
to me, so I wanted to throw that out there.
  
inkwell.vue.505 : Virginia Eubanks, Automating Inequality
permalink #4 of 43: Ari Davidow (ari) Thu 6 Sep 18 11:07
    
Nonetheless, Julie, a good point to consider and a valuable parallel
(imnho). Virginia should be online this evening and I expect things
to start barreling along.
  
inkwell.vue.505 : Virginia Eubanks, Automating Inequality
permalink #5 of 43: Jon Lebkowsky (jonl) Thu 6 Sep 18 14:16
    
Virginia Eubanks was just interviewed for the Plutopia News Network
podcast:
https://www.plutopia.io/2018/09/06/virginia-eubanks-automating-inequality/ Suggest giving a listen as background for this discussion. 

You can share a world readable version of this conversation at
http://bit.ly/automating-inequality

If you're reading this conversation and are not a member of the
WELL, you can still ask a question or contribute a comment via email
to inkwell at well.com.
  
inkwell.vue.505 : Virginia Eubanks, Automating Inequality
permalink #6 of 43: Pamela McCorduck (pamela) Thu 6 Sep 18 15:48
    
Like Ari, I found myself hit harder by this book than I expected.

I was fuming about brittle, ungiving systems that allow no
flexibility on the part of ordinary humans, whether recipients or
officials. I knew that these systems don't *have* to be this way.
With enough initial investment these problems were and are
foreseeable, and when they're not fixable, you go back to parts of
the old system (as the Indiana state system was finally forced to, a
"hybrid" system").  

But the larger truth is that in their rigidity and parsimoniousness,
they reflect exactly the mind-set of the people who ordered up and
implemented them. 

Julie, your point about poorly understood illnesses is well-taken.
  
inkwell.vue.505 : Virginia Eubanks, Automating Inequality
permalink #7 of 43: Virginia Eubanks (veubanks) Thu 6 Sep 18 19:08
    
Hi all! I'm so excited to be here. It means a lot to hear that the
book has resonated with folks emotionally. That was important to me,
and a big reason that I chose for the book to be a work of
journalism rather than academic scholarship. More on that soon,
though... ?

To Ari's question, I usually talk about one particular experience as
the origin moment for the book -- I come from a background in the
community technology and welfare rights/economic justice movements.
In the late 1990s and early 2000s, I worked a lot with a community
of women living in a residential YWCA in my home town of Troy NY. 

I was sitting in a community computer lab one morning with a young
mom on public assistance -- she goes by the pseudonym "Dorothy
Allen" in my first book -- and shooting the breeze about technology.
In the course of that conversation, I asked what she thought about
her EBT (Electronic Benefits Transfer) card, which is the ATM-like
card that now delivers SNAP and cash benefits. They were fairly new
at the time.

She said that it was a little more convenient, but that her
caseworker now had access to her purchase records, and used it to
track all her movements and spending. 

I must have looked really (naively) shocked, because she laughed at
me for a while, and then said, "Oh Virginia! You all [meaning
professional middle-class people] should pay attention to what's
happening to us. You're next!" That was in 2000 - so eighteen years
ago. I think she was remarkably prescient.

Dorothy's voice is always in my head when I begin a writing project
about technology. She reminds me how important it is to start with
people most impacted by new tools. Not only because we should care
about justice and equity, but because they have the best information
about them. 

That old saw by William Gibson, "The future is already here, it's
just not very evenly distributed," is true. But I think I mean it
the opposite way that he meant it. For Gibson, the rich and
privileged get the future first. But when we looked at tools of
digital surveillance and social control, the future comes first to
the marginalized and exploited: poor and working-class communities,
people of color, migrants (especially those who are without legal
status), people who live in war zones, etc.

My goal with the book was to go out and ask poor and working-class
families what their experiences were like as *targets* of new
systems being integrated into social assistance: automated
decisions, matching systems, and predictive analytics in welfare,
homeless services, and child protective services. My instinct was
that the stories they told would be very different from the stories
we tend to hear in the closed circles of technology policy (and even
a lot of technology-focused activism). 
  
inkwell.vue.505 : Virginia Eubanks, Automating Inequality
permalink #8 of 43: Virginia Eubanks (veubanks) Fri 7 Sep 18 05:38
    
And to Julie and Pamela:

Thanks for joining the conversation! You've hit on key pieces of the
argument of the book. 

We often treat new tech as outside of history and context -- like
the monolith in 2001: A Space Odyssey, it just appears from nowhere,
lands on blank ground, and changes everything. But really, tech
grows out of culture and loops back to influence it. 

I spend the first chapter of *Automating Inequality* establishing
that we've had largely punitive, moralistic, and paternalistic
poverty policy in the US, at least since the county poorhouse of the
early 1800s. County poorhouses were brick-and-mortar institutions
for incarcerating poor and working-class people who asked for public
aid. There were more than 1,000 of them across the U.S., and the one
in my home town -- the Rensselaer County House of Industry -- was
open until 1954. In order to enter the poorhouse, you had to give up
your rights to vote and hold office (if you had them), marry, and,
often, you had to give up your kids. The death rates at these
institutions was as high as 30% annually.

What's most important about them for the purpose of the book is that
they represented an important political choice we made in the US,
that our social service system would make the conditions of
receiving support so terrifying that only the most desperate would
ever ask for help. We decided social assistance should be more moral
thermometer -- judging who is worthy and unworthy -- rather than a
universal floor under us all. 

I use the metaphor of the "digital poorhouse" to capture how our new
high-tech tools are part of this history, while still offering new
challenges and possibilities. I think of it as the "deep social
programming" that goes into many of the new tools in social
services.  
  
inkwell.vue.505 : Virginia Eubanks, Automating Inequality
permalink #9 of 43: Ari Davidow (ari) Fri 7 Sep 18 08:29
    
Thank you for your responses, Virginia. One of the reasons the book
resonates is that, unlike, say, "Weapons of Math Destruction" (which
I regard as critical reading on its on terms) you aren't talking as
much about the oversights and lack of critical thinking behind
deploying AI and Big Data systems--you are talking about deliberate
continuity in our spurning of the poor (something that seems to go
back to the the early European settlements on this continent, and is
also at the root of American racism). 

As you write, the issue isn't something new, but rather, new tools
to =automate= inequality. Before we go on to how much of this
affects "the rest of us," can we dig more deeply into what is
happening with the "digital poorhouse"--how it is used to deny
services, but also, as you wrote earlier, how the tracking
mechanisms and the denial of privacy rights may ultimately affect
all of us?
  
inkwell.vue.505 : Virginia Eubanks, Automating Inequality
permalink #10 of 43: Pamela McCorduck (pamela) Fri 7 Sep 18 08:37
    
May? Do already. But we're middle class and not so much at risk. 

That's why I so admire that opening, where Virginia talks about a
medical emergency that comes at a particularly inopportune time (new
job, etc.) and how she, as a well-educated woman with flexible time
from work still is run ragged. It's such a telling incident. 
  
inkwell.vue.505 : Virginia Eubanks, Automating Inequality
permalink #11 of 43: Virtual Sea Monkey (karish) Fri 7 Sep 18 14:27
    
Welcome, Virginia!

I'm struck by the differences and the overlap between what you wrote
about and what Cathy O'Neil wrote about in "Weapons of Math
Distraction". She wrote about errors that grew out of bad technique
in interpreting data, largely by assigning unjustifiable meanings to
the data. You write about the use of technology as a tool and as an
excuse for pursuing antisocial agendas.

You may have gone too easy on the insurance company in your
introduction. Our experience is that they repeatedly use "technical
error" as an excuse to deny benefits, and try to wear down
policyholders even over small benefits.
  
inkwell.vue.505 : Virginia Eubanks, Automating Inequality
permalink #12 of 43: Virginia Eubanks (veubanks) Fri 7 Sep 18 20:40
    
Thanks, Virtual Sea Monkey.

If only I could have validated my hunches about the insurance
company!! I do think that the repeated denial of service when you
are at your most vulnerable -- like when I was caring for my injured
and traumatized partner -- is a pretty clear (and callous) example
of diversion. It's particularly frustrating to have them blame it on
a "technical error" when you are someone who has been studying
technological sleights-of-hand for two decades. 

I absolutely see that if I had even one more challenge, or a little
less support, during that time, I could have easily given up on
getting the health care that we deserved. I was so tired,
demoralized, and vulnerable. I don't think this ended up in the
book, but the day that my employer finally got our health insurance
re-instated, I came down with strep throat. I was bed-ridden for
something like three weeks.

I like *Weapons of Math Destruction* so much. I read it avidly and
refer people to it all the time. It's particularly useful to have a
quant like O'Neil -- an insider -- willing to go on record with her
insights about systemic flaws in these systems. But I wanted to
spend time on a slightly different goal: developing the historical
context and human cost of these tools so that we can understand that
there bad outcomes, even when they are not intentional, are
predictable and avoidable if you take equity and justice seriously. 
  
inkwell.vue.505 : Virginia Eubanks, Automating Inequality
permalink #13 of 43: Virginia Eubanks (veubanks) Sun 9 Sep 18 10:11
    
> Before we go on to how much of this
> affects "the rest of us," can we dig more deeply into what is
> happening with the "digital poorhouse"--how it is used to deny
> services, but also, as you wrote earlier? 

I write in the book about three different data-based systems being
integrated into social service programs across the US: an attempt to
automate and privatize all the eligibility process for welfare
programs in Indiana; an electronic registry of the unhoused in Los
Angeles County that is supposed to match the most vulnerable
homeless with the most appropriate available housing resources; and
a statistical model in Allegheny County, Pennsylvania that is
supposed to predict which children might be victims of abuse or
neglect in the future.

Indiana is a pretty stark example of how (perhaps unconscious)
assumptions about poor and working-class people -- for example, that
they are prone to fraud, lazy, shiftless, addicted, etc -- get baked
into technological design and implementation with devastating
effects. For example, the automated eligibility system interpreted
every mistake in an application as a willful "failure to cooperate
in establishing eligibility," and used errors as reasons to deny
people access to benefits they were eligible for and needed to keep
their families safe and healthy.

Los Angeles raises, among other things, a concern about how the
criminalization of poverty shapes the new tools of poor relief. The
coordinated entry system in LA County was designed and built by very
smart, well-intentioned people who care deeply about the well-being
of the people their agencies serve. But the system is built on top
of a very invasive survey called the VI-SPDAT (Vulnerability Index
and Service Prioritization Decisions Assistance Tool) that asks, for
example, if you are currently trading sex for drugs or if there is
an open warrant on you. And federal data regulations then allow some
of the information coordinated entry collects to be shared with law
enforcement based just on an oral request -- no warrant, no
oversight at all. This leaves many unhoused people feeling that they
have to incriminate themselves in exchange for a slightly better
lottery ticket for housing.

And finally, the Allegheny Family Screening Tool (AFST) raises what
I think might be the book's most troubling question: what does it
mean that we will spend millions of dollars to try and predict how
poor parents might behave in the future, but we won't spend what is
necessary to ensure poor families' basic human rights to shelter,
medical care, and food? For me, in the context of the US' unique
social assistance system, punishment and prediction all are tied up
in each other, leading parents to tell me that they felt like the
prediction tool confuses parenting while poor with poor parenting.  

For each story, I began with the point of view of those people most
impacted by the new tool -- welfare applicants who lost their
benefits during the Indiana automation, unhoused people who have
been surveyed with the VI-SPDAT in Skid Row and South Los Angeles,
and child-welfare involved parents in Allegheny County. I don't stop
there, of course. I did more than a hundred interviews for the book,
and they included designers, data scientists, policy makers,
organizers and activists as well as poor and working families. But
starting with the most impacted allowed me to tell deeper, richer
stories.  
  
inkwell.vue.505 : Virginia Eubanks, Automating Inequality
permalink #14 of 43: paralyzed by a question like that (debunix) Sun 9 Sep 18 17:28
    
>And federal data regulations then allow some
of the information coordinated entry collects to be shared with law
enforcement based just on an oral request -- no warrant, no
oversight at all.

Wow.  
  
inkwell.vue.505 : Virginia Eubanks, Automating Inequality
permalink #15 of 43: Pamela McCorduck (pamela) Mon 10 Sep 18 07:27
    
Each episode had its own horror. Yet I couldn't help feeling, as I
said above, that these outcomes represent some basic societal
hostility to poor people. 

The term "deserving poor" has been with us for a very long time, but
the last forty years of metrics in almost any field have been around
money, which makes anybody poor "undeserving." 
  
inkwell.vue.505 : Virginia Eubanks, Automating Inequality
permalink #16 of 43: Virginia Eubanks (veubanks) Mon 10 Sep 18 10:49
    
Pamela, are you saying that the root of the problem is the social
attitude, not the tech? I get that...but part of the argument I'm
trying to make is that the technology embodies -- and often
amplifies or intensifies -- social attitudes like our fear of
poverty and hatred of the poor.
  
inkwell.vue.505 : Virginia Eubanks, Automating Inequality
permalink #17 of 43: Ari Davidow (ari) Mon 10 Sep 18 16:11
    
Virginia, you last comment also raises obvious questions. First off,
how do we change the use of technology, in the short term, to
dehumanize the poor--is this as simple as getting states to review
procedures? (Obviously not--even when the goals are seemingly
positive, as in Massachusetts' "Health Care Connector," navigating
the system is still a nightmare (and suspect--the system launched
with horrific bugginess).

And longer term, have you had success--has anyone had
success--addressing that fear and hatred of the poor? Does the
problem get a bit easier since the great recession given how much
more widespread the effects of income inequality have become, and
how many more of us know people directly affected?
  
inkwell.vue.505 : Virginia Eubanks, Automating Inequality
permalink #18 of 43: Pamela McCorduck (pamela) Tue 11 Sep 18 07:18
    
Yes, I'm saying the root of the problem is the social attitude. The
tech, as you say, embodies those social attitudes.

Thus the tech itself is ill-planned (good planning takes money and
flexibility to fix when things inevitably go wrong). I had visions
of lowest-bid services getting the contract, and government entities
getting what they paid for. I was not entirely surprised that the
Indiana vendor was IBM, and a bit surprised that the firm prevailed
in court when things blew up.

Your point that technology also amplifies or intensifies social
attitudes is spot on, and I got that too. It's a point well worth
making (shouting from the rooftops).

Big programs always have bugs, sometimes lots of them. Payrolls,
which we've been doing for more than half a century with computers,
are finally pretty stable. But note the modifier: "pretty." That's
why Ronald Reagan's Brilliant Pebbles/aka Star Wars, was so
hopeless. It had to work right the first time. That's why our
anti-ballistic missile programs are so piss-poor in their
performances. (Those are different sides of the
amplification/intensification phenomenon.)

The homeless program in Los Angeles seemed to me a slightly
different issue. People really wanted to be helpful to the homeless.
How could they have done it better?
  
inkwell.vue.505 : Virginia Eubanks, Automating Inequality
permalink #19 of 43: Betsy Schwartz (betsys) Tue 11 Sep 18 08:20
    
Part of the problem in LA was  , as someone in the book says, a
carpentry problem. All the computer-matching in the world could not
compensate for the loss of what was it, 10,000+ units?

And in Indiana it seemed that, beyond all the bugs, the technology
was doing what it was designed to do: reducing the headcount and
cost of public services. 

Victoria, thank you for joining us and I'll agree it was a brutally
hard book to read, but important. 

It was very interesting to read how this traced back decades, and
about the inherent contradiction between wanting to support the
genuinely needy while not making it tempting for the lazy. Somehow,
today, this contradiction manifests as actively penalizes people who
are receiving assistance for trying in any way to improve their
situation. And the massively intrusive surveillance enforces this. 
  
inkwell.vue.505 : Virginia Eubanks, Automating Inequality
permalink #20 of 43: Jon Lebkowsky (jonl) Wed 12 Sep 18 16:35
    
My career working with technology began when I was working
eligibility for poverty programs (Food Stamps, AFDC, Medicaid) in
the 80s and 90s. Texas had developed an electronic version of the
worksheet we used for eligibility determination. It worked
surprisingly well given the complexity of the regulations we were
implementing. Our system was pretty objective. 

Virginia wrote about Indiana's concern that caseworkers might
collude with their clients - this was a driver for greater reliance
on automation. In Texas, the caseworker's job didn't really change,
the automation was created to support the eligibility determination,
but the worker's judgement was still involved. However we didn't see
collusion with clients - in fact some caseworkers had more of a
police mentality, looking patterns suggesting fraud. 

I think they implemented a new system years after I was gone, and
that it was more like Indiana's.

Getting back to complexity of the regulations - the impact of that
complexity was often underestimated, and it's effect on operations
was sometimes interpreted from the outside as inefficiency. I recall
that a team of developers was brought in at some point, confident
that they could build a new, more efficient version of the system.
In the beginning they were confident, but as they continued working
their confidence waned - they realized that the tool we had
developed was actually about as good as it was going to get.

From what Virginia has written, I think the problem in Indiana was
probably attributable to mismanagement, and to in correct
assumptions and expectations on the part of the outside vendors. In
Texas, our software was developed by internal teams working with
subject matter experts, and the result was quite functional and
useable.
  
inkwell.vue.505 : Virginia Eubanks, Automating Inequality
permalink #21 of 43: Virtual Sea Monkey (karish) Wed 12 Sep 18 16:54
    
From what Virginia wrote about Indiana I see mismanagement,
unrealistic expectations, and bad faith.
  
inkwell.vue.505 : Virginia Eubanks, Automating Inequality
permalink #22 of 43: Ari Davidow (ari) Thu 13 Sep 18 08:55
    
I think "mismanagement" is too kind. The point is that we think that
there is something wrong with people who require aid, which leads us
to use these systems to further prune eligibility. 

We reinforce the animus against the poor by also assaulting their
privacy. If you are poor, you are forced to submit to a significant
loss of privacy. If you use a card to purchase food, the government
knows what you are purchasing. But, more generally, the applications
to get aid require you to reveal personal details that would never
be required of people not requiring government aid--they are clearly
far too intrusive.

This all leads to my big question: How do we change attitudes (and
aid is one part of this--criminal justice reform is another major
place where we are likelier to use big data to perpetuate
inequality, rather than improve outcomes)? How do we change our
perspective from "they're all cheats, so cut benefits, lock them up,
deny services" to "when people need help, we need to find ways to
ensure that they have the help they need"?

As Utah Phillips used to complain, we're good at attacking the minor
miscreants, and all-too-eager to let those "job creator" folks who
have done well in our system pay a fraction of their fair share.
That may well be far beyond the scope of this particular book, but
it is very much on my mind.

Getting back to the book, Virginia, what are the outcomes you were
looking for when you wrote the book? More awareness? Public
discussion on how we "automate inequality" and what that means?
Specific changes to programs and legislation?
  
inkwell.vue.505 : Virginia Eubanks, Automating Inequality
permalink #23 of 43: Jon Lebkowsky (jonl) Fri 14 Sep 18 12:09
    
Ari, in the Texas system at the time I was working there, there was
an avoidance of judgemental bias, and we were also taking privacy
seriously. I can't say whether that's been the case in the many
years since I left, but my point is that the technology itself had
been developed avoiding "animus against the poor." I guess I'm
saying that the use of technology in and of itself is not the
problem.
  
inkwell.vue.505 : Virginia Eubanks, Automating Inequality
permalink #24 of 43: Virginia Eubanks (veubanks) Fri 14 Sep 18 15:23
    
I'm deeply suspicious of the "it's not the technology, it's the
people" line of argument because technology IS people. It is
designed and implemented by humans and holds human values and
decisions. I think it is dangerous to separate human values from
their manifestation in technology. It's like the snake biting its
tail -- technology emerges out of a certain set of values and
doubles back to shape and reshape society. 

So often, people try to establish the neutrality of tech by arguing
that it is "just a tool." One of the odder jobs I hold is as a brick
mason. And I have at least SIX different trowels. I'm barely an
amateur -- I mostly repoint and repair historic brick -- so I have a
very limited kit. But every one of my tools is essential for a
different task, and uniquely shaped to it. I can't use a 1/4 inch
tuck pointing trowel to smooth an expanse of concrete, for example.
I can't even use a 1/4 inch tuck pointing trowel to repoint a
building with narrower or wider joints. 

What I'm trying to say is that there is nothing neutral or blank
about tools. They evolve over time to fit a specific purpose and
while you can use a hammer to paint a barn, you'll do a terrible
job.

Similarly, I believe that the deep social programming of our newest
high-tech tools -- rooted as they are in hundreds of years of
punitive and criminalizing poverty policy in the US -- makes them
ill-suited to social justice. It's not impossible to bend them
towards better outcomes -- just like it's not impossible to paint a
barn with a hammer. But it is really, really hard.
  
inkwell.vue.505 : Virginia Eubanks, Automating Inequality
permalink #25 of 43: Ari Davidow (ari) Sat 15 Sep 18 06:36
    
There can be a fine line, I think, between the thoughtless
interfaces of, say, common online forms--filled in a credit card
number or phone number online lately?--and the numbing, bewildering
experience of the systems we have put in place of people as the
front line (or only line) to provide access to needed services. But,
one is stupid and inconvenient; the latter is used to winnow out all
but the best at filling out online forms and denies needed services
to those who don't succeed. If we saw that help as a necessary part
of who we are as a society, we would be using different tools in
different ways.
  

More...



Members: Enter the conference to participate. All posts made in this conference are world-readable.

Subscribe to an RSS 2.0 feed of new responses in this topic RSS feed of new responses

 
   Join Us
 
Home | Learn About | Conferences | Member Pages | Mail | Store | Services & Help | Password | Join Us

Twitter G+ Facebook