Back to overview
Categories
Uncategorized

Using Language Sciences for Social Good

Language technologies, like ChatGPT, are developing quickly. But we don’t really know how they work. And they are not always designed to help society.

12 juni 2024, Nikki Weststeijn

Illustration by Arco Mul

If ChatGPT gives us an answer to a factual question, we have no idea if this answer is really true. Open AI, the makers of ChatGPT, bear no responsibility that their chatbot and text generator will reply truthfully. Similarly, if an employer uses some type of language technology to screen the CV’s of applicants, there is no promise that the technology will not be biased in some way.

This is because these technologies are all self-learning algorithms. It is a type of artificial intelligence that is trained by seeing sets of exemplary inputs and correct outputs and then learns some strategy to come up with outputs on its own. It is unclear why the algorithm will give one answer and not the other.

Prof dr Floris Roelofsen

Floris Roelofsen, professor at the ILLC, is director of a new large-scale research project called ‘Language Sciences for Social Good’ (LSG) at the University of Amsterdam. The idea behind this overarching research project is to make sure that language technology will also be used for social good, and not just guided by cost-efficiency and result-driven goals. Roelofsen: “Language technology is booming, it’s really transforming society. With ChatGPT it’s very clear, but this development has been going on for the past 25 years. We have been using language technology when we search on google, when we translate pieces of text, or when we use voice assistants on our phone.”

Language technology has a lot of impact, but in areas where it could potentially have a positive impact, it is still lacking, according to Roelofsen. The LSG project has a threefold goal: to use more responsible methods in developing language technology, and make sure that language technology can contribute to a safe society and an inclusive society.

One of the main goals in terms of using responsible methods in language sciences, is that language technology should be more transparent. Going back to the CV-screening example: in the training phase, the algorithm is shown sets of CV’s and the cv that in the past has actually been picked by an employer. The algorithm then learns a way to select a CV out of a pile of CV’s, the one that is supposed to belong to the best candidate for the job.

Unless we add additional rules to the program, this algorithm will not make the decision in a way that is similar to how a human makes the decision. It is not a decision based on certain principles, like the fact that the candidate needs to be trustworthy, hardworking and smart. It will just be a decision based on what is the most likely outcome given the examples that it has seen.

A researcher at the ILLC who has started doing research related to the transparency of AI is Fausto Carcassi. He is a cognitive scientist and is working on a model that uses probabilities to determine the meaning of a word in different contexts. Carcassi hopes that methods of this model can also be used by Large Language Models (LLM’s) such as ChatGPT and that, vice versa, techniques used by LLM’s can also improve the model he is working on.

Dr Fausto Carcassi
Fausto Carcassi: Creating transparent AI by improving linguistic understanding

Carcassi’s research is an example of how creating more transparent AI could also contribute to developing safer AI. Safety of AI is a big concern. Language technology can be harmful on different levels. It can offend people, for example by producing a sexist poem. If used to screen cv’s, it could discriminate certain social groups, for example by picking out white men more often. A chatbot like ChatGPT can also be directly endangering, for example by providing instructions on how to build a bomb upon request.

Ekaterina Shutova, Associate Professor of Natural Language Processing at the ILLC, does research on how we can prevent technologies like ChatGPT from producing hate speech or otherwise discriminatory content. She recently received a grant from Meta, the company that owns Facebook and Instagram, to do similar research on misinformation.

Dr Ekaterina Shutova
Ekaterina Shutova: How to prevent AI from making hateful, biased and misinformed comments

Shutova is working on a method to take out bias áfter the model is already designed and trained. This is because, at the moment, most companies that develop AI still consider safety to be an issue that should be dealt with after the technology is developed. However, AI can in principle also be developed with safety as one of the main requirements.

One could even develop AI with positive goals in mind, such as to make society more inclusive. Roelofsen: “The development of language technology is now mostly driven by commercial interests, even though there is so much potential to design things that can be good for society. Universities are not commercial. And I believe that it would be a waste if we would be doing the same things commercial companies are already doing.”

An example of ways in which language technology can contribute to a more inclusive society is the sign language project that Roelofsen has been working on for a few years. One of the aims of this project is to create an avatar that can produce sign language. This is much needed according to Roelofsen, because currently deaf people are in many ways excluded from society. Roelofsen has a young daughter who is born deaf and has noticed that Dutch sign language (NGT) is not really taken seriously as a language. “Currently, in the Netherlands the consensus seems to be that when a child is born deaf, it should learn to speak Dutch. But this is much harder when you cannot perceive the sounds. It’s like having to learn Chinese while never being able to actually listen to it.”

According to Roelofsen, spoken Dutch will always be like a second language to deaf people. It costs way more effort to understand, even if it is written down. Since spoken Dutch is used everywhere, it can be harder for deaf people to participate in society. This is why Roelofsen wants to create an avatar that can communicate with deaf people in their own language, for example in public spaces such as train stations. As part of this sign language project, Jos Ritmeester, a linguist at the ILLC, is working on an app that can help parents of deaf children learn sign language.

Jos Ritmeester
Jos Ritmeester: An app that can help parents of deaf children learn sign language

Ritmeester’s project is an example of one of the ways in which language technology can help people who speak a minority language feel more included in society. In general, the large language models that are developed now, are trained mostly on English texts. Simply because these models need lots of data and English texts are widely available. These models cannot speak languages that are used by fewer people in the world. Such as sign languages for example.

But it is exactly for these types of languages that language technology could really play an important role. For immigrants in the Netherlands who do not speak Dutch, for example, it is much harder to decide who to vote for, to understand government programs like the vaccination programs during the Covid pandemic, or to defend themselves in court. Language technology could help these people to become more included in society. Roelofsen: “It can really make a difference for some people.”