“It is difficult to create a universal ethical model for machines.”

Xabier Saralegi Urizar is an artificial intelligence researcher at Orai NLP Technologies, a new technological initiative promoted by the Elhuyar Foundation. It tries to use the full potential of digitalization in favor of Basque. His team researches automatic translators, voice synthesis, and other technologies. It also aims to combat the discriminatory and racist behaviour of these technologies. In short, the creation of more ethical and human machines.


Machines are very good at arithmetic. Not so much communicating with humans. But they also do it better and better. Do speech technologies have a lot of potential?

Yeah, they got it. In the field of artificial intelligence there are two disciplines that, although very complex, have gained great strength: on the one hand, vision – recognition and image processing – and, on the other, natural language processing. Very complex algorithms are needed to automate this process that the human brain naturally performs in machines. For example: machine translation, question and answer systems, text summarization systems, grammatical correctors... But we are getting better and better results.

Do you want machines to communicate with humans?

Yeah, but it's not just that. After all, most of the information that is generated in the world is encoded in our language. This big data exists in human language and in digital form. Therefore, this source is the most productive to access any information. In order to make the most of the information we have digitized and to carry out a variety of analyses, natural language processing is necessary. This is how we will benefit from this information.

But you've never said that machines are racist and exclusionary.

Yeah, that's right. In order for machines to learn the language of humans, they are trained with texts created by humans: news, novels... These texts look at our behaviors; they are a reflection of our racist and exclusionary society. Therefore, machines learn exclusionary behaviors.

In machine translation, for example, it is seen very clearly in the professions: often the engineer considers them directly as men and the nurse considers them as women. The same was true of religious minorities: a system that extracted synonyms or semantic relations, for example, sometimes linked Islam to terrorism. Google also developed a service to detect objects in photos, which in some cases classified black people as gorillas.

How can you show ethics to a machine?

In reality, bias or exclusionary attitudes come from training texts that we need to “clean up”. The fact is that there are many minorities and many ethical conflicts that sometimes intersect. It is not feasible to create millions of purified texts and a universal ethical corpus that has no bias in any social problem. It’s ethically difficult, not just technologically.

Ethical models can be created by paying attention to a specific problem, and we are working on it. But the algorithms we create for machine learning can also amplify the bias inherent in texts. Both factors must be taken into account.

I'm talking about Arg. By Jon Urbe/Photo Press
If machines want to communicate and understand us with humans, they will need to detect and understand our emotions. Is this possible?

Now, mainly neural language models are used for machines to learn natural language, and in these neural language models they try to distinguish between different degrees of abstraction of language: morphological, syntactic, semantic and pragmatic. The latter is where emotions come in. A scale of six categories has been proposed to detect emotions: angry, happy, sad... What happens is that in spoken language, gestures and intonation must also be taken into account to detect emotions. Therefore, image processing is also required. It requires a neural architecture that combines three disciplines: image recognition, speech technology and language comprehension. That's what they're already investigating.

It's harder to detect the irony. To do this, the system should pay attention to the whole and the global context, what we call “knowledge of the world”. To understand jokes, the machine must first learn some cultural references. Integrating all this knowledge of the world into neural networks is very complex. There's no neural network that makes good jokes, for example. But that will come, too.

You talk about neural networks, but how do they look like neurons?

For example, the automatic translator Elia that we have developed, when we want to translate a phrase, must take a phrase and announce how it is said in another language. It therefore has two variables: input and output. A word can be introductory, and some neurons have to predict how to translate it. These neurons operate with bits: they receive a number at the input and give a number at the output. In binary language, 0 or 1.

I mean, what you call a neuron is really a mathematical formula?

Yes, a formula. You have a series of values in the input, and in order to provide the output value, the input values are combined. But for every input word, you need a bunch of interconnected neurons. And because of the complexity of language, millions of neurons, millions of mathematical formulas, interconnected, are ultimately needed. Like biological neurons. Hence the name of the neural networks. These neural architectures can be organized in layers. In a certain way, morphology is learned in the first layer; in the second, syntax; in the third, semantics

And in order to get better results, we need to adjust these formulas. But how do you adjust a million formulas? It cannot be done by hand. We use optimization algorithms until we explore automatically and find the parameters that give the best results in training.

How did machines learn before the invention of neural networks?

It has had a long evolution. In the beginning, methodologies based on language rules were used: “If in a sentence this word appears in this position, and has a pronoun next to it...” But the language is complex and there are too many rules, so the results were not good. Then they thought that statistical models would be the most appropriate. The results improved slightly. in 2010, Mikolov, a Google researcher, proposed a technique to better represent the linguistic characteristics of words. And they started using deep neural networks.

I'm talking about Arg. By Jon Urbe/©Photo Press

But in 2017 and 2018, two innovations were proposed by other Google researchers to better learn the complexity of language. These innovations have consolidated today’s paradigm: the knowledge of a language is concentrated in a huge neural linguistic model that can be adapted to perform specific tasks of language processing. This has led to a boom in language technologies.

Since machines don’t have the ability to abstract, they need a lot of examples to learn anything. Does this make it difficult to create technology for small languages?

Yeah, it's a big limit. We are working on a project with the HiTZ centre and Vicomtech to obtain the largest database that the Basque language has ever had. we've reached a corpus of 400 million words. In English, they drive billions.

Today, the health of languages is also measured in the digital sphere. Therefore, language policies should take into account the development of technology, right?

It is clear that the most important thing to strengthen the digital vitality of a language is the creation of content. But then we need digital tools to generate interaction in this language: automatic translators, automatic subtitling, news recommendations... To consume content, language technologies are necessary.

Communication has been completely transformed by digitalization. At this time, you can read the tweet written by a Russian journalist in your own language. But for this, languages need technological support. Otherwise, they will be out of the game in the short term. So, yes, it is important that language policies have a clear digital strategy. The Russian, however, does not have a very clear strategy.

In the interview you mentioned that the main advances came from Google. In this context, do you see technological sovereignty as important?

In my opinion, it is essential to be sovereign in technologies that are critical to the survival of our language. Above all, to be able to mark strategic lines and not depend on anyone. Google offers some services – translators, search engines, transcription systems, etc. – and I often listen to people in fear: “If Google doesn’t come out, we’re out!” But perhaps the most important thing is not that Google does it, but that we acquire the knowledge to do it ourselves. In fact, to have this knowledge is to be sovereign. This will give us the ability to deal with any situation without relying on large corporations.

Most of the innovations have been made by Google researchers, but they have made algorithms and other resources available to everyone. In the Basque Country we have benefited greatly from these resources. We have three teams: HiTZ Center, Vicomtech and Orai NLP Technologies. So we have the ingredients to achieve technological sovereignty. But strong public funding is needed, because the market will never save the weak.

Buletina

Bidali zure helbide elektronikoa eta jaso asteroko buletina zure sarrera-ontzian

Bidali

Bizitza