technology bias

Machined Prejudice: Three Sources of Technology Bias

Given our long history with tools, the idea that we inject bias into technology isn’t exactly new. What is new is the way that machine learning introduces subtle new forms of technology bias.

Technology Bias: the embedding of a particular tendency, trend, inclination, feeling, or opinion into technological systems

1) Designers and Technology Bias

The most obvious ways we bias our tools is through the assumptions we bring to the design process. Sometimes those assumptions are deliberate, but more often than not, they are unconscious.

silent_house_partyThe other day, my wife and I came across a guy blasting a mini boombox as he walked down the street in our neighborhood. As a society, we’ve now collectively opted for designs that direct audio streams solely into the ears of an individual listener. It’s a logical approach, but as each of us move into our own private audioscape, we gradually erode community (and enable some unusual partying behavior).

All design decisions are judgments, and as such, convey some form of bias. We often just don’t notice it — especially the tool has been with us for awhile.

2) End Users and Technology Bias

screen-shot-2016-10-09-at-12-12-40-pmDigital media enables us to interact with information in new ways. That feedback loop introduces a new form of technology bias. As end users now participate in creating products and services, they introduce bias through their engagement.

The way we like and share stuff on social media streams, for example, doesn’t just shape our own experience. It also influences what happens to our friends on these networks. Your bias for cute kittens, clever memes and birthday messages increases my likelihood of seeing that stuff in my stream. Our interactions with each another cause the network to become our bias.

The strange thing about end user bias is that radically different types of bias can coexist simultaneously on the same platform. Clusters of hatred and bigotry can thrive right beside communities of love and inspiration. Our engagement fragments us into echo chambers of shared bias.

3) Algorithm Trainers and Technology Bias

Machine learning algorithms learn by interacting with humans, often via services like Google Search and Facebook. What that means is that we humans are training the artificial intelligence that fuels our intelligent devices. What that also means is that human trainers play an important role in determining the values — and biases — of these systems.

image-identificationLet’s say, for example, that I want to build a machine learning algorithm that recognize images. Let’s also say that to train the system, I select a group of trainers who are all men. The human intelligence passed into this system is thereby skewed towards a male perspective. It might be so biased as to annoy, and possibly even insult, women who to tried to use it.

Training bias is a serious concern. In machine learning, selecting human trainers is a core part of the design process. As more of these systems come online, learning and growing through their interactions with us, we must guard against imbuing them with harmful human bias.

unprofessional-hairstylesMachine learning now grades essays on standardized tests and automates resume screening in HR systems. It’s important that these systems aren’t trained in ways that favor certain groups of people over others. Police departments across the U.S. are already assessing people’s likelihood of committing future crimes based on a system with a demonstrated racial bias. In a society already wrestling with institutional racism, a thoughtless rush to artificial intelligence could replicate bias on an unimaginable scale that could unravel the very fabric of society.

Designing Containers of Culture

Tribunal-AwakeningsI believe that artificial intelligence is becoming a container for collective human intelligence. This question of technology bias shows how artificial intelligence also acts as a container for human culture. We’re still in the early days of defining what this container will look like, but Riot Games provides an intriguing hint in its use of machine learning to change the toxic culture of online gaming.

As intelligent systems control more and more aspects of society and our economy, it’s essential that we learn to identify and isolate harmful bias as a proactive part of the design process for any intelligent system. Doing so won’t just weaken the grip of frail human egos. It will strengthen the better angels of our culture.

 

 

 

Silent house party by Imokurnotok  CC BY-SA 3.0,

2 comments

  1. One person likes a boom-box. Others like iPhones. Where is the bias? Is it “bias”, that other people are hearing one person’s music? How?
    How do you propose to “correct” the opinions of humans without introducing your own bias?

    • The bias in that example is really just a designer responding to customer preferences. It’s kind of a silly example, actually. I’m using the boombox, which is something I’d not seen someone using on the street in a very long time, to show how most designers no longer build that product with that use in mind – though they might have a couple decades ago. It’s a design assumption, or bias, that isn’t really consciously stated anywhere. It just happens.

      As for correcting people’s opinions, that’s not what this is about. It’s about building systems with a conscious eye towards the fact that bias is going to creep in through a variety of ways and building options into our services to give people a way of removing those filter biases, if they they so choose.

Leave a Reply