Measuring biases of GenAI听

In her latest research, Dr. Afra Mashhadi and fellow researchers seek to align generative AI with human understanding.听

In today鈥檚 technology-driven world, we鈥檙e constantly inundated with new tools designed to improve our quality of life and maximize efficiency. From the way we brush our teeth in the morning to how we fall asleep at night, there鈥檚 an app, smart appliance or other digital service purporting to help us do it better. 

While these tools may offer certain benefits, it鈥檚 not without the potential to cause harm. Perhaps no technology today is more hotly debated for its risk versus utility than artificial intelligence. Despite the endless possible positive applications of AI, the risks can include everything from intellectual property theft and misinformation to its threats to jobs and the environment. 

The potential for bias is another challenge that AI presents 鈥 one that Dr. Afra Mashhadi, assistant professor in the 糖心vlog视频鈥檚 School of STEM, has made the central focus in her latest research. 

Guardians of technology

A researcher and educator, Mashhadi is passionate about advocating for ethics and diversity in computing. Her research focuses on computational behavioral modeling and ethical AI, with a particular interest in generative AI. 

GenAI, or AI systems capable of generating new content, has made significant strides in recent years and has exploded in popularity in the form of widely used tools such as ChatGPT, a chatbot that uses large language models to generate human-like text responses. The rapid speed with which the technology continues to evolve and expand, however, often leaves ethical checks and balances in the dust, Mashhadi said. 

鈥淧eople need to know whether these tools they鈥檙e using are biased, whether they鈥檙e trustworthy or not,鈥 she said. 鈥淭he companies that are creating these models, they鈥檙e not going to be the ones that look at the safety of them. I think universities have a powerful and important role to play, which cannot be substituted by industry where they have different priorities. 

鈥淚n the university system, we are like the guardians of technology,鈥 she noted. 鈥淚t鈥檚 for us to examine and audit these models to understand how we can make them better.鈥 

In her latest research 鈥 鈥淎uditLab: Algorithm and Tools for Measuring Biases of GenAI Models鈥 鈥 Mashhadi collaborated with Dr. Chirag Shah, a professor at the UW鈥檚 , and a team of student researchers to create a set of algorithms and methods for evaluating how large language models perform in social reasoning domains. 

Collaborative research

The AuditLab project was funded through UW Bothell鈥檚 Scholarship, Research and Creative Practice Grant Program, which provided financial support for student researchers and other resources. It was also made possible through the collaboration across schools, Mashhadi added, as the iSchool provided the infrastructure and dedicated server access needed to run larger models. 

Mashhadi, who is also an adjunct professor at the iSchool, brought students from both schools onto the project. 

鈥淚 came to the Computer Science program at UW Bothell primarily to go deeper into natural language processing, so work in this area was a career goal from the start,鈥 said Maryna Sivachenko, (Computer Science & Software Engineering 鈥25). 鈥淚 was extremely happy that I was able to join this project to examine what鈥檚 behind this tool that people use but don鈥檛 understand how it鈥檚 performing, what the pitfalls and limitations are, and how can we do better right now.鈥 

In the early research for the project, the group focused mainly on demographic biases, such as gender and ethnicity. What they found was that the two large language models they examined both tended to produce more accurate links in co-authorship networks for individuals with Asian or white names, particularly among researchers with lower visibility or limited academic impact. The team recently in the European Physical Journal Data Science. 

As they continued their research, their focus shifted to another aspect of potential bias 鈥 a misalignment in understanding human language and behavior. 

“We as humans are the gatekeepers of this AI output, and there鈥檚 a lot of room to grow this research.”

Elias Martin, MS in Computer Science & Software Engineering 鈥25听

Theory of mind

鈥淲e don鈥檛 know how much natural language understanding these LLMs have, so our motivation was to explore that alignment,鈥 said Maryam Amirizaniani, a doctoral student at the iSchool. 鈥淭here are many different applications of LLMs, and if industry and even academia are going to use them as a service provider for customer service or answering mental health questions or educational tutoring, they need to have a human understanding.鈥 

To evaluate how closely LLMs align with human understanding, the team looked to the theory of mind, which refers to the capacity for humans to recognize that people have different beliefs, desires, intentions and emotions from one another. They discovered that the LLMs often fell short in their understanding when compared to human responses to open-ended questions on platforms such as Lemmy and Reddit. 

Looking to help close the gap, the team again looked to the theory of mind by creating a method for identifying relevant information about the user鈥檚 mental state to feed back into the models. 

鈥淲e collected data from social forums like Reddit and Lemmy and tried to extract some implicit information that exists in the user鈥檚 questions 鈥 like their emotions and the intentions they have behind the question and the sentiment that exists 鈥 and then gave it to the LLMs as main points for consideration,鈥 Amirizaniani said. 鈥淲hen we applied this framework, their understanding about humans improved, and they could generate more human-like answers.鈥 

The team presented its findings at two major conferences and has additional papers on their way to being published. 

鈥淔or me, as someone who hadn鈥檛 done research like this before, it was super exciting to have this option and to be the only person at the conference at a bachelor鈥檚 level,鈥 Sivachenko said. 鈥淚 really loved this reasoning alignment topic because of my curiosity about the natural language understanding of machines and how they can be better or worse depending on certain factors.鈥 

A seed planted

While the project findings provide valuable information for helping to improve generative AI, there鈥檚 still a lot of research to be done. One of the biggest challenges to the field is how quickly it continues to advance, said Elias Martin, who graduated with his master鈥檚 in Computer Science & Software Engineering from UW Bothell in June. 

鈥淎s these large language models and other AI keep getting added to new contexts and applied in new ways, new biases are surfacing every day,鈥 he said. 鈥淚t鈥檚 really challenging to keep up with how much data is being generated and effectively review it all, but ultimately, we as humans are the gatekeepers of this AI output, and there鈥檚 a lot of room to grow this research.鈥 

Mashhadi鈥檚 students all noted an interest in continuing to do research in this area, and they centered their capstone projects around the topic as well. 

鈥淚鈥檓 really honored that my students who were on this project decided to work on related topics on their capstone. I鈥檇 like to think that maybe the spark came from this project,鈥 she said. 

鈥淭he SRCP is a seed grant that really helped us to plant the beginnings of this new research. Now we can see how much more we need to do. It鈥檚 been amazing to see how teaching can fit into the pipeline of research.鈥 

Read more recent news

See all news