Featured Product
This Week in Quality Digest Live
Innovation Features
Jill Roberts
Another way to know what’s too old to eat
Gregory Way
Drug designers are rethinking their response to medications that affect multiple targets
Del Williams
8-in. cable and disc systems are comparable to belt or bucket systems
Edmund Andrews
For creative collaboration, sometimes you can’t beat a face-to-face meeting
Steven Brown
21st-century standard candles at NIST

More Features

Innovation News

More News

Innovation

Building Machines That Work for Everyone

How diversity of test subjects is a technology blind spot, and what to do about it

Published: Wednesday, February 2, 2022 - 13:02

People interact with machines in countless ways every day. In some cases, they actively control a device, like driving a car or using an app on a smartphone. Sometimes people passively interact with a device, like being imaged by an MRI machine. And sometimes they interact with machines without consent or even knowing about the interaction, like being scanned by a law enforcement facial-recognition system.

Human-machine interaction (HMI) is an umbrella term that describes the ways people interact with machines. It’s a key aspect of researching, designing, and building new technologies as well as also studying how people use and are affected by technologies.

Researchers, especially those traditionally trained in engineering, increasingly are taking a human-centered approach when developing systems and devices. This means striving to make technology that works as expected for the people who will use it, by taking into account what’s known about the people and by testing the technology with them. But even as engineering researchers increasingly prioritize these considerations, some in the field have a blind spot: diversity.

As an interdisciplinary researcher who thinks holistically about engineering and design, and an expert in dynamics and smart materials with interests in policy, we have examined the lack of inclusion in technology design, the negative consequences, and possible solutions.

People at hand

Researchers and developers typically follow a design process of testing key functions and features before releasing products to the public. Done properly, the process can be a key component of compassionate design. The tests can include interviews and experiments with groups of people who stand in for the public.

In academic settings, for example, most study participants are students. Some researchers attempt to recruit off-campus participants, but these communities are often similar to the university population. Coffee shops and other locally owned businesses, for example, may allow fliers to be posted in their establishments. However, the clientele of these establishments is often students, faculty, and academic staff.

In many industries, co-workers serve as test participants for early-stage work because it’s convenient to recruit from within a company. It takes effort to bring in outside participants, and when they’re used, they often reflect the majority population. Therefore, many of the people who participate in these studies have similar demographic characteristics.

Real-world harm

It’s possible to use a homogenous sample of people when publishing a research paper that adds to a field’s body of knowledge. And some researchers who conduct studies this way acknowledge the limitations of homogenous study populations. However, when it comes to developing systems that rely on algorithms, such oversights can cause real-world problems. Algorithms are as only as good as the data that are used to build them.

Algorithms are often based on mathematical models that capture patterns and then inform a computer about those patterns to perform a given task. Imagine an algorithm designed to detect when colors appear on a clear surface. If the set of images used to train that algorithm consists of mostly shades of red, the algorithm might not detect when a shade of blue or yellow is present.

In practice, algorithms have failed to: detect darker skin tones for Google’s skincare program and in automatic soap dispensers; accurately identify a suspect, which led to the wrongful arrest of an innocent man in Detroit; and reliably identify women of color. MIT artificial intelligence researcher Joy Buolamwini describes this as algorithmic bias and has extensively discussed and published work on these issues.

Even as the United States fights Covid-19, the lack of diverse training data has become evident in medical devices. Pulse oximeters, which are essential for keeping track of your health at home and to indicate when you might need hospitalization, may be less accurate for people with melanated skin. These design flaws, like those in algorithms, are not inherent to the device but can be traced back to the technology being designed and tested using populations that weren’t diverse enough to represent all potential users.

Being inclusive

Researchers in academia are often under pressure to publish their findings as quickly as possible. Therefore, reliance on convenience samples—that is, people who are easy to reach and get data from—is common.

Though institutional review boards exist to ensure that study participants’ rights are protected and that researchers follow proper ethics in their work, they don’t have the responsibility to dictate to researchers whom they should recruit. When researchers are pressed for time, considering different populations for study subjects can mean additional delay. Finally, some researchers may simply be unaware of how to adequately diversify their study’s subjects.

There are several ways researchers in academia and industry can increase the diversity of their study participant pools.

One is to make time to do the inconvenient and sometimes hard work of developing inclusive recruitment strategies. This can require creative thinking. One such method is to recruit diverse students who can serve as ambassadors to diverse communities. The students can gain research experience while also serving as a bridge between their communities and researchers.

Another is to allow members of the community to participate in the research and provide consent for new and unfamiliar technologies whenever possible. For example, research teams can form an advisory board composed of members from various communities. Some fields frequently include an advisory board as part of their government-funded research plans.

Another approach is to include people who know how to think through cultural implications of technologies as members of the research team. For instance, the New York City Police Department’s use of a robotic dog in Brooklyn, Queens, and the Bronx sparked outrage among residents. This might have been avoided if they had engaged with experts in the social sciences or science and technology studies, or simply consulted community leaders.

Lastly, diversity is not just about race but also age, gender identity, cultural backgrounds, educational levels, disability, English proficiency, and even socioeconomic levels. Lyft is on a mission to deploy robotaxis next year, and experts are excited about the prospects of using robotaxis to transport the elderly and disabled. It’s not clear whether these aspirations include those who live in less-affluent or low-income communities, or lack the family support that could help prepare people to use the service. Before dispatching a robotaxi to transport grandmothers, it’s important to take into account how a diverse range of people will experience the technology.The Conversation

This article is republished from The Conversation under a Creative Commons license. Read the original article.

Discuss

About The Authors

Tahira Reid’s picture

Tahira Reid

Tahira Reid is an associate professor of mechanical engineering at Purdue University.

James Gibert’s picture

James Gibert

James Gibert is anssociate professor of mechanical engineering at Purdue University.