Artificial intelligence can discriminate on the basis of race and gender, and also age – Tech Xplore


Forget Password?
Learn more
share this!
10
23
Share
Email
January 19, 2022
by Charlene Chu, Kathleen Leslie, Rune Nyrup, Shehroz Khan,
We have accepted the use of artificial intelligence (AI) in complex processes—from health care to our daily use of social media—often without critical investigation, until it is too late. The use of AI is inescapable in our modern society, and it may perpetuate discrimination without its users being aware of any prejudice. When health-care providers rely on biased technology, there are real and harmful impacts.

This became clear recently when a study showed that pulse oximeters—which measure the amount of oxygen in the blood and have been an essential tool for clinical management of COVID-19—are less accurate on people with darker skin than lighter skin. The findings resulted in a sweeping racial bias review now underway, in an attempt to create international standards for testing medical devices.
There are examples in , business, government and everyday life where biased algorithms have led to problems, like sexist searches and racist predictions of an offender’s likelihood of re-offending.
AI is often assumed to be more objective than humans. In reality, however, AI algorithms make decisions based on human-annotated data, which can be biased and exclusionary. Current research on in AI focuses mainly on gender and race. But what about age-related bias—can AI be ageist?
Ageist technologies?
In 2021, the World Health Organization released a global report on aging, which called for urgent action to combat ageism because of its widespread impacts on health and well-being.
Ageism is defined as “a process of systematic stereotyping of and discrimination against people because they are old.” It can be explicit or implicit, and can take the form of negative attitudes, discriminatory activities, or institutional practices.
The pervasiveness of ageism has been brought to the forefront throughout the COVID-19 pandemic. Older adults have been labeled as “burdens to societies,” and in some jurisdictions, age has been used as the sole criterion for lifesaving treatments.
Digital ageism exists when age-based bias and discrimination are created or supported by technology. A recent report indicates that a “” of more than 2.5 quintillion bytes of data is produced each day. Yet even though are using in greater numbers—and benefiting from that use—they continue to be the age cohort least likely to have access to a computer and the internet.
Digital ageism can arise when ageist attitudes influence technology design, or when ageism makes it more difficult for older adults to access and enjoy the full benefits of digital technologies.
Cycles of injustice
There are several intertwined cycles of injustice where technological, individual and social biases interact to produce, reinforce and contribute to digital ageism.
Barriers to technological access can exclude older adults from the research, design and development process of digital technologies. Their absence in technology design and development may also be rationalized with the ageist belief that older adults are incapable of using technology. As such, older adults and their perspectives are rarely involved in the development of AI and related policies, funding and support services.
The unique experiences and needs of older adults are overlooked, despite age being a more powerful predictor of technology use than other demographic characteristics including race and gender.
AI is trained by data, and the absence of older adults could reproduce or even amplify the above ageist assumptions in its output. Many AI technologies are focused on a stereotypical image of an older adult in poor health—a narrow segment of the population that ignores healthy aging. This creates a that not only discourages older adults from using AI, but also results in further data loss from these demographics that would improve AI accuracy.
Even when older adults are included in large datasets, they are often grouped according to arbitrary divisions by developers. For example, older adults may be defined as everyone aged 50 and older, despite younger age cohorts being divided into narrower age ranges. As a result, older adults and their needs can become invisible to AI systems.
In this way, AI systems reinforce inequality and magnify societal exclusion for sections of the population, creating a “digital underclass” primarily made up of older, poor, racialized and marginalized groups.
Addressing digital ageism
We must understand the risks and harms associated with age-related biases as more older adults turn to technology.
The first step is for researchers and developers to acknowledge the existence of digital ageism alongside other forms of algorithmic biases, such as racism and sexism. They need to direct efforts towards identifying and measuring it. The next step is to develop safeguards for AI systems to mitigate ageist outcomes.
There is currently very little training, auditing or oversight of AI-driven activities from a regulatory or legal perspective. For instance, Canada’s current AI regulatory regime is sorely lacking.
This presents a challenge, but also an opportunity to include ageism alongside other forms of biases and discrimination in need of excision. To combat digital ageism, older adults must be included in a meaningful and collaborative way in designing new technologies.
With bias in AI now recognized as a critical problem in need of , it is time to consider the experience of digital for older adults, and understand how growing old in an increasingly digital world may reinforce social inequalities, exclusion and marginalization.


Explore further

Effects of digital technologies on older people’s access to health and social care


Provided by The Conversation

Explore further
This article is republished from The Conversation under a Creative Commons license. Read the original article.The Conversation
Facebook
Twitter
Email
Feedback to editors
1 hour ago
0
23 hours ago
0
Jan 18, 2022
1
Jan 18, 2022
0
Jan 18, 2022
1
1 hour ago
1 hour ago
23 hours ago
23 hours ago
Jan 19, 2022
Jan 19, 2022
Jan 19, 2022
Dec 02, 2021
Jun 09, 2016
Jul 13, 2020
Nov 19, 2021
Jun 21, 2019
Jan 15, 2020
Jan 19, 2022
Jan 19, 2022
Jan 18, 2022
Jan 18, 2022
Jan 18, 2022
Jan 17, 2022
Use this form if you have come across a typo, inaccuracy or would like to send an edit request for the content on this page. For general inquiries, please use our contact form. For general feedback, use the public comments section below (please adhere to guidelines).
Please select the most appropriate category to facilitate processing of your request
Thank you for taking time to provide your feedback to the editors.
Your feedback is important to us. However, we do not guarantee individual replies due to the high volume of messages.
Your email address is used only to let the recipient know who sent the email. Neither your address nor the recipient’s address will be used for any other purpose. The information you enter will appear in your e-mail message and is not retained by Tech Xplore in any form.

Daily science news on research developments and the latest scientific innovations
Medical research advances and health news
The most comprehensive sci-tech news coverage on the web
This site uses cookies to assist with navigation, analyse your use of our services, collect data for ads personalisation and provide content from third parties. By using our site, you acknowledge that you have read and understand our Privacy Policy and Terms of Use.

source
Connect with Chris Hood, a digital strategist that can help you with AI.

Leave a Reply

Your email address will not be published.

© 2022 AI Caosuo - Proudly powered by theme Octo