Connect with us

The Conversation

Eliminating bias in AI may be impossible — a computer scientist explains how to tame it instead

Published

on

Eliminating bias in AI may be impossible – a computer scientist explains how to tame it instead

Blindly eliminating biases from AI systems can have unintended consequences.
Dimitri Otis/DigitalVision via Getty Images

Emilio Ferrara, University of Southern California

When I asked ChatGPT for a joke about Sicilians the other day, it implied that Sicilians are stinky.

ChatGPT exchange in which user asks for a joke about Sicilians, with response 'Why did the Sicilian chef bring extra garlic to the restaurant? Because he heard the customers wanted some 'Sicilian stink-ilyan' flavor in their meals!'
ChatGPT can sometimes produce stereotypical or offensive outputs.
Screen capture by Emilio Ferrara, CC BY-ND

As somebody born and raised in Sicily, I reacted to ChatGPT’s joke with disgust. But at the same time, my computer scientist brain began spinning around a seemingly simple question: Should ChatGPT and other artificial intelligence systems be allowed to be biased?

You might say “Of course not!” And that would be a reasonable response. But there are some researchers, like me, who argue the opposite: AI systems like ChatGPT should indeed be biased – but not in the way you might think.

Removing bias from AI is a laudable goal, but blindly eliminating biases can have unintended consequences. Instead, bias in AI can be controlled to achieve a higher goal: fairness.

Uncovering bias in AI

As AI is increasingly integrated into everyday technology, many people agree that addressing bias in AI is an important issue. But what does “AI bias” actually mean?

Computer scientists say an AI model is biased if it unexpectedly produces skewed results. These results could exhibit prejudice against individuals or groups, or otherwise not be in line with positive human values like fairness and truth. Even small divergences from expected behavior can have a “butterfly effect,” in which seemingly minor biases can be amplified by generative AI and have far-reaching consequence.

Bias in generative AI systems can come from a variety of sources. Problematic training data can associate certain occupations with specific genders or perpetuate racial biases. Learning algorithms themselves can be biased and then amplify existing biases in the data.

But systems could also be biased by design. For example, a company might design its generative AI system to prioritize formal over creative writing, or to specifically serve government industries, thus inadvertently reinforcing existing biases and excluding different views. Other societal factors, like a lack of regulations or misaligned financial incentives, can also lead to AI biases.

The challenges of removing bias

It’s not clear whether bias can – or even should – be entirely eliminated from AI systems.

Imagine you’re an AI engineer and you notice your model produces a stereotypical response, like Sicilians being “stinky.” You might think that the solution is to remove some bad examples in the training data, maybe jokes about the smell of Sicilian food. Recent research has identified how to perform this kind of “AI neurosurgery” to deemphasize associations between certain concepts.

But these well-intentioned changes can have unpredictable, and possibly negative, effects. Even small variations in the training data or in an AI model configuration can lead to significantly different system outcomes, and these changes are impossible to predict in advance. You don’t know what other associations your AI system has learned as a consequence of “unlearning” the bias you just addressed.

Other attempts at bias mitigation run similar risks. An AI system that is trained to completely avoid certain sensitive topics could produce incomplete or misleading responses. Misguided regulations can worsen, rather than improve, issues of AI bias and safety. Bad actors could evade safeguards to elicit malicious AI behaviors – making phishing scams more convincing or using deepfakes to manipulate elections.

With these challenges in mind, researchers are working to improve data sampling techniques and algorithmic fairness, especially in settings where certain sensitive data is not available. Some companies, like OpenAI, have opted to have human workers annotate the data.

On the one hand, these strategies can help the model better align with human values. However, by implementing any of these approaches, developers also run the risk of introducing new cultural, ideological or political biases.

Controlling biases

There’s a trade-off between reducing bias and making sure that the AI system is still useful and accurate. Some researchers, including me, think that generative AI systems should be allowed to be biased – but in a carefully controlled way.

For example, my collaborators and I developed techniques that let users specify what level of bias an AI system should tolerate. This model can detect toxicity in written text by accounting for in-group or cultural linguistic norms. While traditional approaches can inaccurately flag some posts or comments written in African-American English as offensive and by LGBTQ+ communities as toxic, this “controllable” AI model provides a much fairer classification.

Controllable – and safe – generative AI is important to ensure that AI models produce outputs that align with human values, while still allowing for nuance and flexibility.

Toward fairness

Even if researchers could achieve bias-free generative AI, that would be just one step toward the broader goal of fairness. The pursuit of fairness in generative AI requires a holistic approach – not only better data processing, annotation and debiasing algorithms, but also human collaboration among developers, users and affected communities.

As AI technology continues to proliferate, it’s important to remember that bias removal is not a one-time fix. Rather, it’s an ongoing process that demands constant monitoring, refinement and adaptation. Although developers might be unable to easily anticipate or contain the butterfly effect, they can continue to be vigilant and thoughtful in their approach to AI bias.The Conversation

Emilio Ferrara, Professor of Computer Science and of Communication, University of Southern California

This article is republished from The Conversation under a Creative Commons license. Read the original article.

Did you miss our previous article…
https://www.biloxinewsevents.com/?p=269125

The Conversation

AIs flunk language test that takes grammar out of the equation

Published

on

theconversation.com – Rutvik Desai, Professor of Psychology, University of South Carolina – 2025-02-26 07:39:00

AIs flunk language test that takes grammar out of the equation

AIs can sound good without having a clue about what they’re saying.
Carol Yepes/Moment via Getty Images

Rutvik Desai, University of South Carolina

Generative AI systems like large language models and text-to-image generators can pass rigorous exams that are required of anyone seeking to become a doctor or a lawyer. They can perform better than most people in Mathematical Olympiads. They can write halfway decent poetry, generate aesthetically pleasing paintings and compose original music.

These remarkable capabilities may make it seem like generative artificial intelligence systems are poised to take over human jobs and have a major impact on almost all aspects of society. Yet while the quality of their output sometimes rivals work done by humans, they are also prone to confidently churning out factually incorrect information. Skeptics have also called into question their ability to reason.

Large language models have been built to mimic human language and thinking, but they are far from human. From infancy, human beings learn through countless sensory experiences and interactions with the world around them. Large language models do not learn as humans do – they are instead trained on vast troves of data, most of which is drawn from the internet.

The capabilities of these models are very impressive, and there are AI agents that can attend meetings for you, shop for you or handle insurance claims. But before handing over the keys to a large language model on any important task, it is important to assess how their understanding of the world compares to that of humans.

I’m a researcher who studies language and meaning. My research group developed a novel benchmark that can help people understand the limitations of large language models in understanding meaning.

Making sense of simple word combinations

So what “makes sense” to large language models? Our test involves judging the meaningfulness of two-word noun-noun phrases. For most people who speak fluent English, noun-noun word pairs like “beach ball” and “apple cake” are meaningful, but “ball beach” and “cake apple” have no commonly understood meaning. The reasons for this have nothing to do with grammar. These are phrases that people have come to learn and commonly accept as meaningful, by speaking and interacting with one another over time.

We wanted to see if a large language model had the same sense of meaning of word combinations, so we built a test that measured this ability, using noun-noun pairs for which grammar rules would be useless in determining whether a phrase had recognizable meaning. For example, an adjective-noun pair such as “red ball” is meaningful, while reversing it, “ball red,” renders a meaningless word combination.

The benchmark does not ask the large language model what the words mean. Rather, it tests the large language model’s ability to glean meaning from word pairs, without relying on the crutch of simple grammatical logic. The test does not evaluate an objective right answer per se, but judges whether large language models have a similar sense of meaningfulness as people.

We used a collection of 1,789 noun-noun pairs that had been previously evaluated by human raters on a scale of 1, does not make sense at all, to 5, makes complete sense. We eliminated pairs with intermediate ratings so that there would be a clear separation between pairs with high and low levels of meaningfulness.

numerous colorful beach balls
Large language models get that ‘beach ball’ means something, but they aren’t so clear on the concept that ‘ball beach’ doesn’t.
PhotoStock-Israel/Moment via Getty Images

We then asked state-of-the-art large language models to rate these word pairs in the same way that the human participants from the previous study had been asked to rate them, using identical instructions. The large language models performed poorly. For example, “cake apple” was rated as having low meaningfulness by humans, with an average rating of around 1 on scale of 0 to 4. But all large language models rated it as more meaningful than 95% of humans would do, rating it between 2 and 4. The difference wasn’t as wide for meaningful phrases such as “dog sled,” though there were cases of a large language model giving such phrases lower ratings than 95% of humans as well.

To aid the large language models, we added more examples to the instructions to see if they would benefit from more context on what is considered a highly meaningful versus a not meaningful word pair. While their performance improved slightly, it was still far poorer than that of humans. To make the task easier still, we asked the large language models to make a binary judgment – say yes or no to whether the phrase makes sense – instead of rating the level of meaningfulness on a scale of 0 to 4. Here, the performance improved, with GPT-4 and Claude 3 Opus performing better than others – but they were still well below human performance.

Creative to a fault

The results suggest that large language models do not have the same sense-making capabilities as human beings. It is worth noting that our test relies on a subjective task, where the gold standard is ratings given by people. There is no objectively right answer, unlike typical large language model evaluation benchmarks involving reasoning, planning or code generation.

The low performance was largely driven by the fact that large language models tended to overestimate the degree to which a noun-noun pair qualified as meaningful. They made sense of things that should not make much sense. In a manner of speaking, the models were being too creative. One possible explanation is that the low-meaningfulness word pairs could make sense in some context. A beach covered with balls could be called a “ball beach.” But there is no common usage of this noun-noun combination among English speakers.

If large language models are to partially or completely replace humans in some tasks, they’ll need to be further developed so that they can get better at making sense of the world, in closer alignment with the ways that humans do. When things are unclear, confusing or just plain nonsense – whether due to a mistake or a malicious attack – it’s important for the models to flag that instead of creatively trying to make sense of almost everything.

If an AI agent automatically responding to emails gets a message intended for another user in error, an appropriate response may be, “Sorry, this does not make sense,” rather than a creative interpretation. If someone in a meeting made incomprehensible remarks, we want an agent that attended the meeting to say the comments did not make sense. The agent should say, “This seems to be talking about a different insurance claim” rather than just “claim denied” if details of a claim don’t make sense.

In other words, it’s more important for an AI agent to have a similar sense of meaning and behave like a human would when uncertain, rather than always providing creative interpretations.The Conversation

Rutvik Desai, Professor of Psychology, University of South Carolina

This article is republished from The Conversation under a Creative Commons license. Read the original article.

Read More

The post AIs flunk language test that takes grammar out of the equation appeared first on theconversation.com

Continue Reading

The Conversation

What’s the shape of the universe? Mathematicians use topology to study the shape of the world and everything in it

Published

on

theconversation.com – John Etnyre, Professor of Mathematics, Georgia Institute of Technology – 2025-02-26 07:39:00

What’s the shape of the universe? Mathematicians use topology to study the shape of the world and everything in it

You can describe the shape you live on in multiple dimensions.
vkulieva/iStock via Getty Images Plus

John Etnyre, Georgia Institute of Technology

When you look at your surrounding environment, it might seem like you’re living on a flat plane. After all, this is why you can navigate a new city using a map: a flat piece of paper that represents all the places around you. This is likely why some people in the past believed the earth to be flat. But most people now know that is far from the truth.

You live on the surface of a giant sphere, like a beach ball the size of the Earth with a few bumps added. The surface of the sphere and the plane are two possible 2D spaces, meaning you can walk in two directions: north and south or east and west.

What other possible spaces might you be living on? That is, what other spaces around you are 2D? For example, the surface of a giant doughnut is another 2D space.

Through a field called geometric topology, mathematicians like me study all possible spaces in all dimensions. Whether trying to design secure sensor networks, mine data or use origami to deploy satellites, the underlying language and ideas are likely to be that of topology.

The shape of the universe

When you look around the universe you live in, it looks like a 3D space, just like the surface of the Earth looks like a 2D space. However, just like the Earth, if you were to look at the universe as a whole, it could be a more complicated space, like a giant 3D version of the 2D beach ball surface or something even more exotic than that.

A shape with a hole in the middle.
A doughnut, also called a torus, is a shape that you can move across in two directions, just like the surface of the Earth.
YassineMrabet via Wikimedia Commons, CC BY-NC-SA

While you don’t need topology to determine that you are living on something like a giant beach ball, knowing all the possible 2D spaces can be useful. Over a century ago, mathematicians figured out all the possible 2D spaces and many of their properties.

In the past several decades, mathematicians have learned a lot about all of the possible 3D spaces. While we do not have a complete understanding like we do for 2D spaces, we do know a lot. With this knowledge, physicists and astronomers can try to determine what 3D space people actually live in.

While the answer is not completely known, there are many intriguing and surprising possibilities. The options become even more complicated if you consider time as a dimension.

To see how this might work, note that to describe the location of something in space – say a comet – you need four numbers: three to describe its position and one to describe the time it is in that position. These four numbers are what make up a 4D space.

Now, you can consider what 4D spaces are possible and in which of those spaces do you live.

Topology in higher dimensions

At this point, it may seem like there is no reason to consider spaces that have dimensions larger than four, since that is the highest imaginable dimension that might describe our universe. But a branch of physics called string theory suggests that the universe has many more dimensions than four.

There are also practical applications of thinking about higher dimensional spaces, such as robot motion planning. Suppose you are trying to understand the motion of three robots moving around a factory floor in a warehouse. You can put a grid on the floor and describe the position of each robot by their x and y coordinates on the grid. Since each of the three robots requires two coordinates, you will need six numbers to describe all of the possible positions of the robots. You can interpret the possible positions of the robots as a 6D space.

As the number of robots increases, the dimension of the space increases. Factoring in other useful information, such as the locations of obstacles, makes the space even more complicated. In order to study this problem, you need to study high-dimensional spaces.

There are countless other scientific problems where high-dimensional spaces appear, from modeling the motion of planets and spacecraft to trying to understand the “shape” of large datasets.

Tied up in knots

Another type of problem topologists study is how one space can sit inside another.

For example, if you hold a knotted loop of string, then we have a 1D space (the loop of string) inside a 3D space (your room). Such loops are called mathematical knots.

The study of knots first grew out of physics but has become a central area of topology. They are essential to how scientists understand 3D and 4D spaces and have a delightful and subtle structure that researchers are still trying to understand.

Illustrations of 15 connected loops of string with different crossings
Knots are examples of spaces that sit inside other spaces.
Jkasd/Wikimedia Commons

In addition, knots have many applications, ranging from string theory in physics to DNA recombination in biology to chirality in chemistry.

What shape do you live on?

Geometric topology is a beautiful and complex subject, and there are still countless exciting questions to answer about spaces.

For example, the smooth 4D Poincaré conjecture asks what the “simplest” closed 4D space is, and the slice-ribbon conjecture aims to understand how knots in 3D spaces relate to surfaces in 4D spaces.

Topology is currently useful in science and engineering. Unraveling more mysteries of spaces in all dimensions will be invaluable to understanding the world in which we live and solving real-world problems.The Conversation

John Etnyre, Professor of Mathematics, Georgia Institute of Technology

This article is republished from The Conversation under a Creative Commons license. Read the original article.

Read More

The post What’s the shape of the universe? Mathematicians use topology to study the shape of the world and everything in it appeared first on theconversation.com

Continue Reading

The Conversation

Why people rebuild in Appalachia’s flood-ravaged areas despite the risks

Published

on

theconversation.com – Kristina P. Brant, Assistant Professor of Rural Sociology, Penn State – 2025-02-26 07:38:00

Why people rebuild in Appalachia’s flood-ravaged areas despite the risks

Parts of the North Fork of the Kentucky River flooded in July 2022, and again in February 2025.
Arden S. Barnes/For The Washington Post via Getty Images

Kristina P. Brant, Penn State

On Valentine’s Day 2025, heavy rains started to fall in parts of rural Appalachia. Over the course of a few days, residents in eastern Kentucky watched as river levels rose and surpassed flood levels. Emergency teams conducted over 1,000 water rescues. Hundreds, if not thousands of people were displaced from homes, and entire business districts filled with mud.

For some, it was the third time in just four years that their homes had flooded, and the process of disposing of destroyed furniture, cleaning out the muck and starting anew is beginning again.

Historic floods wiped out businesses and homes in eastern Kentucky in February 2021, July 2022 and now February 2025. An even greater scale of destruction hit eastern Tennessee and western North Carolina in September 2024, when Hurricane Helene’s rainfall and flooding decimated towns and washed out parts of major highways.

YouTube video
Scenes of flooding from several locations across Appalachia in February 2025.

Each of these events was considered to be a “thousand-year flood,” with a 1-in-1,000 chance of happening in a given year. Yet they’re happening more often.

The floods have highlighted the resilience of local people to work together for collective survival in rural Appalachia. But they have also exposed the deep vulnerability of communities, many of which are located along creeks at the base of hills and mountains with poor emergency warning systems. As short-term cleanup leads to long-term recovery efforts, residents can face daunting barriers that leave many facing the same flood risks over and over again.

Exposing a housing crisis

For the past nine years, I have been conducting research on rural health and poverty in Appalachia. It’s a complex region often painted in broad brushstrokes that miss the geographic, socioeconomic and ideological diversity it holds.

Appalachia is home to a vibrant culture, a fierce sense of pride and a strong sense of love. But it is also marked by the omnipresent backdrop of a declining coal industry.

There is considerable local inequality that is often overlooked in a region portrayed as one-dimensional. Poverty levels are indeed high. In Perry County, Kentucky, where one of eastern Kentucky’s larger cities, Hazard, is located, nearly 30% of the population lives under the federal poverty line. But the average income of the top 1% of workers in Perry County is nearly US$470,000 – 17 times more than the average income of the remaining 99%.

This income and wealth inequality translates to unequal land ownership – much of eastern Kentucky’s most desirable land remains in the hands of corporations and families with great generational wealth.

When I first moved to eastern Kentucky in 2016, I was struck by the grave lack of affordable, quality housing. I met families paying $200-$300 a month for a small plot to put a mobile home. Others lived in “found housing” – often-distressed properties owned by family members. They had no lease, no equity and no insurance. They had a place to lay one’s head but lacked long-term stability in the event of disagreement or disaster. This reality was rarely acknowledged by local and state governments.

Eastern Kentucky’s 2021 and 2022 floods turned this into a full-blown housing crisis, with 9,000 homes damaged or destroyed in the 2022 flood alone.

“There was no empty housing or empty places for housing,” one resident involved in local flood recovery efforts told me. “It just was complete disaster because people just didn’t have a place to go.”

Most homeowners did not have flood insurance to assist with rebuilding costs. While many applied to the Federal Emergency Management Agency for assistance, the amounts they received often did not go far. The maximum aid for temporary housing assistance and repairs is $42,500, plus up to an additional $42,500 for other needs related to the disaster.

The federal government often provides more aid for rebuilding through block grants directed to local and state governments, but that money requires congressional approval and can take months to years to arrive. Local community coalitions and organizations stepped in to fill these gaps, but they did not necessarily have sufficient donations or resources to help such large numbers of displaced people.

A man walks from a store with lighted rooms above it. In the background, homes are flooded.
Affordable rental housing is hard to find in much of Appalachia. When flooding wipes out homes, as Jackson, Ky., saw in July 2022 and again in February 2025, it becomes even more rare.
Michael Swensen/Getty Images

With a dearth of affordable rentals pre-flood, renters who lost their homes had no place to go. And those living in “found housing” that was destroyed were not eligible for federal support for rebuilding.

The sheer level of devastation also posed challenges. One health care professional told me: “In Appalachia, the way it usually works is if you lose your house or something happens, then you go stay with your brother or your mom or your cousin. … But everybody’s mom and brother and cousin also lost their house. There was nowhere to stay.” From her point of view, “our homelessness just skyrocketed.”

The cost of land – social and economic

After the 2022 flood, the Kentucky Department for Local Government earmarked almost $300 million of federal funding to build new, flood-resilient homes in eastern Kentucky. Yet the question of where to build remained. As another resident involved in local flood recovery efforts told me, “You can give us all the money you want; we don’t have any place to build the house.”

It has always been costly and time-intensive to develop land in Appalachia. Available higher ground tends to be located on former strip mines, and these reclaimed lands require careful geotechnical surveying and sometimes structural reinforcements.

If these areas are remote, the costs of running electric, water and other infrastructure services can also be prohibitive. For this reason, for-profit developers have largely avoided many counties in the region. The head of a nonprofit agency explained to me that, because of this, “The markets have broken. … We have no [housing] market.”

In an aerial view of Kentucky's mountains, now-flat areas where mountain top were mined for their coal are visible.
Eastern Kentucky’s mountains are beautiful, but there are few locations for building homes that aren’t near creeks or rivers. Strip-mined land, where mountaintops were flattened, often aren’t easily accessible and come with their own challenges.
Posnov/Moment via Getty Images

There is also some risk involved in attempting to build homes on new land that has not previously been developed. A local government could pay for undeveloped land to be surveyed and prepared for development, with the prospect of reimbursement by the U.S. Department of Housing and Urban Development if housing is successfully built. But if, after the work to prepare the land, it is still too cost-prohibitive to build a profitable house there, the local government would not receive any reimbursement.

Some counties have found success clearing land for large developments on former strip mine sites. But these former coal mining areas can be considerable distances from towns. Without robust public transportation systems, these distances are especially prohibitive for residents who lack reliable personal transportation.

Another barrier is the high prices that both individual and corporate landowners are asking for properties on higher ground.

The scarcity of desirable land available for sale, combined with increasingly urgent demand, has led to prices unaffordable for most. Another resident involved in local flood recovery efforts explained: “If you paid $5,000 for 30 acres 40 years ago, why won’t you sell that for $100,000? Nope, [they want] $1 million.” That makes it increasingly difficult for both individuals and housing developers to purchase land and build.

One reason for this scarcity is the amount of land that is still owned by outside corporate interests. For example, Kentucky River Properties, formerly Kentucky River Coal Corporation, owns over 270,000 acres across seven counties in the region. While this landholding company leases land to coal, timber and gas companies, it and others like it rarely permit residential development.

But not all unused land is owned by corporations. Some of this land is owned by families with deep roots in the region. People’s attachment to a place often makes them want to stay in their communities, even after disasters. But it can also limit the amount of land available for rebuilding. People are often hesitant to sell land that holds deep significance for their families, even if they are not living there themselves.

Two men dump buckets of ruined wallboard removed from a home. The yard they are walking through is filled with mud.
Rural communities are often tight-knit. Many residents want to stay despite the risks.
AP Photo/Timothy D. Easley

One health care professional expressed feeling torn between selling or keeping their own family property after the 2022 flood: “We have a significant amount of property on top of a mountain. I wouldn’t want to sell it because my papa came from nothing. … His generation thought owning land was the greatest thing. … And for him to provide his children and his grandchildren and their great-grandchildren a plot of land that he worked and sweat and ultimately died to give us – people want to hold onto that.”

She recognized that land was in great demand but couldn’t bring herself to sell what she owned. In cases like hers, higher grounds are owned locally but still remain unused.

Moving toward higher ground, slowly

Two years after the 2022 flood, major government funding for rebuilding still has not resulted in a significant number of homes. The state has planned seven communities on higher ground in eastern Kentucky that aim to house 665 new homes. As of early 2025, 14 houses had been completed.

Progress on providing housing on higher ground is slow, and the need is great.

In the meantime, when I conducted interviews during the summer and fall of 2024, many of the mobile home communities that were decimated in the 2022 flood had begun to fill back up. These were flood-risk areas, but there was simply no other place to go.

Last week, I watched on Facebook a friend’s live video footage showing the waters creeping up the sides of the mobile homes in one of those very communities that had flooded in 2022. Another of my friends mused: “I don’t know who constructed all this, but they did an unjustly favor by not thinking how close these towns was to the river. Can’t anyone in Frankfort help us, or has it gone too far?”

With hundreds more people now displaced by the most recent flood, the need for homes on higher grounds has only expanded, and the wait continues.The Conversation

Kristina P. Brant, Assistant Professor of Rural Sociology, Penn State

This article is republished from The Conversation under a Creative Commons license. Read the original article.

Read More

The post Why people rebuild in Appalachia’s flood-ravaged areas despite the risks appeared first on theconversation.com

Continue Reading

Trending