Feature

Gender is personal – not computational

Efforts at automatic gender recognition – using algorithms to guess a person’s gender based on images, video or audio – raise significant social and ethical concerns that are not yet fully explored.

Gender

Should an algorithm try to guess what gender people are by how they look? Source: all_is_magic/Shutterstock.com

Imagine walking down the street and seeing advertising screens change their content based on how you walk, how you talk, or even the shape of your chest. These screens rely on hidden cameras, microphones and computers to guess if you’re male or female. This might sound futuristic, but patrons in a Norwegian pizzeria discovered it’s exactly what was happening: Women were seeing ads for salad and men were seeing ads for meat options. The software running a digital advertising board spilled the beans when it . The motivation behind using this technology might have been to improve advertising quality or user experience. Nevertheless, many customers were unpleasantly surprised by it.

This sort of situation is not just creepy and invasive. It’s worse: Efforts at – using algorithms to guess a person’s gender based on images, video or audio – raise significant social and ethical concerns that are not yet fully explored. Most current research on automatic gender recognition technologies focuses instead on technological details.

found that people with diverse gender identities, including those identifying as transgender or gender nonbinary, are particularly concerned that . People who express their gender differently from stereotypical male and female norms already as a result of . Ideally, technology designers should develop systems to make these problems less common, not more so.

Using algorithms to classify people

As digital technologies become more powerful and sophisticated, their designers are trying to use them to identify and categorise , such as . The idea is that with enough , algorithms can learn to analyse people’s appearance and behavior – and perhaps one day characterise people as well as, or even better than, other humans do.
How machine learning works.

Gender is a hard topic for people to handle. It’s a complex concept with important roles both as a cultural construct and a core aspect of an individual’s identity. Researchers, scholars and activists are increasingly revealing the . In the process, they find that ignoring this diversity can lead to both . For example, according to the , 47 percent of transgender participants stated that they had experienced some form of discrimination at their workplace due to their gender identity. More than half of transgender people who were harassed, assaulted or expelled because of their gender identity had attempted suicide.

Many people have, at one time or another, been surprised, or confused or even angered to find themselves mistaken for a person of another gender. When that happens to someone who is transgender – as an estimated , are – it can cause .

Effects of automatic gender recognition

In our recent research, we , about their general impressions of automatic gender recognition technology. We also asked them to describe their responses to imaginary future scenarios where they might encounter it. All 13 participants were worried about this technology and doubted whether it could offer their community any benefits.

Of particular concern was the prospect of being misgendered by it; in their experience, gender is largely an internal, subjective characteristic, not something that is necessarily or entirely expressed outwardly. Therefore, neither humans nor algorithms can accurately read gender through physical features, such as the face, body or voice.

They described how being misgendered by algorithms could potentially feel worse than if humans did it. Technology is often perceived or believed to be , so being wrongly categorised by an algorithm would emphasise the misconception that a transgender identity is inauthentic. One participant described how they would feel hurt if a “million-dollar piece of software developed by however many people” decided that they are not who they themselves believe they are.

Privacy and transparency

The people we interviewed shared the that automated cameras could be used for surveillance without their consent or knowledge; for years, researchers and activists have about increasing threats to privacy in a world populated by sensors and cameras.
file-20180502-153873-fw7qo8.jpg?ixlib=rb-1.1.0&q=45&auto=format&w=754&fit=clip
Facial recognition software can scan a crowd of people as they walk by.

But our participants described how the effects of these technologies could be greater for transgender people. For instance, they might be singled out as unusual because they look or behave differently from what the underlying algorithms expect. Some participants were even concerned that systems might falsely determine that they are trying to be someone else and deceive the system.

Their concerns also extended to cisgender people who might look or act differently from the majority, such as people of different races, people the algorithms perceive as androgynous, and people with unique facial structures. This already happens to people from minority racial and ethnic backgrounds, who are regularly . For example, existing facial recognition technology in some cameras fail to properly detect the faces of Asian users and send messages for them to .

Our interviewees wanted to know more about how automatic gender recognition systems work and what they’re used for. They didn’t want to know deep technical details, but did want to make sure the technology would not compromise their privacy or identity. They also wanted more transgender people involved in the early stages of design and development of these systems, well before they are deployed.

Creating inclusive automatic systems

Our results demonstrate how designers of automatic categorisation technologies can inadvertently cause harm by making assumptions about the simplicity and predictability of human characteristics. Our research adds to a that attempts to more into technology.

Minorities have historically been left out of conversations about large-scale technology deployment, including . Yet, scientists and designers alike know that including input from minority groups during the design process can lead to technical innovations that . We advocate for a more gender-inclusive and human-centric approach to automation that incorporates diverse perspectives.
count.gif?distributor=republish-lightbox-basic
As digital technologies develop and mature, they can lead to impressive innovations. But as humans direct that work, they should that are negative and limiting. In the case of automatic gender recognition, we do not necessarily conclude that these algorithms should be abandoned. Rather, designers of these systems should be inclusive of, and sensitive to, the diversity and complexity of human identity.


, Postdoctoral Research Associate in Information Systems, ; , Master's student in Human-Centered Computing, , and , Lecturer of Information Systems,

This article was originally published on . Read the .


Share
7 min read
Published 16 May 2018 12:29pm
Source: The Conversation


Share this with family and friends