Grzegorz J. Nalepa is a professor at the AGH University in Krakow, Poland who recently held a series of lectures at the Faculty of Computer Science. One of them was “Affective Computing, Context and Processing”. We interviewed him on the ethical questions posed by this new development in technology.

You are a computer scientist but, at the same time, you have a master’s degree in Philosophy. What were the reasons for such an unusual choice?
Let’s say that I decided to approach Philosophy for intellectual pleasure, not for research reasons. When I completed my Master in Computer Science and Artificial Intelligence, I felt that, in some way, I needed to dedicate myself to something more profound, so I took up that study while attending a PhD in Computer Science.

During your stay at unibz, one of your scheduled lectures was on Affective Computing. What is it?
It’s a very broad topic and not entirely a new one. As a matter of fact, almost twenty years ago it was Rosalind Picard - the founder and director of the Affective Computing Research Group at the Massachusetts Institute of Technology - who proposed it for the first time. I think the simplest definition could be that of a domain where, on the one hand, computers are able to simulate and express emotions and, on the other hand, they can recognize and interpret human feelings. These are the two main aspects on which the research focuses nowadays. I work mostly on this second aspect.

Why should someone want computers to detect and interpret or simulate emotions? What’s the use of it?
From a business prospective, there is a huge, open market for such applications. Many companies are interested in the development of technologies that allow the recognition of feelings. Some of them are already doing it. We can think of the “like” buttons on Facebook, for example, which have expanded from the original thumbs up to other emotions.

Prof. Grzegorz J. Nalepa is an expert in Affective Computing

What kind of applications do you refer to?
There are many different ones. For instance, in Computer Science we refer to “Sentiment Analysis“. It’s basically an analysis of a written text on Facebook, Amazon, a website or wherever. Fundamentally, you have a text, a computer and an algorithm: you run the algorithm on the text, whichshould tell you if it’s emotionally positive or negative. More advanced methods could describe a full range of different feelings. Recently there has been a lot of work on the automatic detection of hate speech, for example. Neuromarketing has a great interest in knowing the emotional state of a consumer in order to use this information for different marketing purposes. We could be getting different types of advertising based on our feelings - depending on whether we are sad or happy, we could get a suitable message. In general, marketing will be benefiting from the research in this area.

What about the second dimension of affective computing?
There could be many applications in care systems. These other applications are mainly oriented towards senior citizens. These do not necessarily have to be robots, maybe it could mean an affective communication system between a machine and a user. In this case we want the experience to be as friendly as possible and the system to be either sympathetic or compassionate. If the system can discover how we feel, maybe it could use a better tone and cheer us up!

Clearly there are important debates on possible ethical consequences of AI, rather than specifically on affective computing. But I don’t think such discussions should just be held among us computer scientists.

Grzegorz J. Nalepa

With such gross interference in our emotional world, are we not running the risk of constantly being manipulated by the market?
In general I agree that we’re too observed and monitored. There might be technologies that are invented for evil purposes but I’d say that most technologies are neutral. We can use affective computing in a positive way but, of course, we could use this information to mislead people.

What kind of ethical problems does this pose to a Computer Scientist?

We are, of course, aware that this whole domain could have a significant impact on our societies but scientists are rarely responsible for creating the final products for the market. It’s a long chain that involves people from business, politics and, finally, everyone of us. The problem is that society should understand how the technology works. But this is a general question that goes beyond mere Computer Science. It’s a problem that regards society as a whole. The same applies to cameras we have in the cities, to the monitoring of credit cards, smart phones, and other aspects of our everyday life and technology. Recently I’ve been reading and hearing that people are more and more worried about the development of Artificial Intelligence (AI). But I don’t think we need to be frightened. Fear is always a powerful tool on a commercial and a political level. We need to understand the potential of AI and cope with it; the same applies to Affective Computing.

Is there a debate on this topic in the community of computer scientists?
Clearly there are important debates on possible ethical consequences of AI, rather than specifically on affective computing. But I don’t think such debates should just be held among us computer scientists. In some countries there are foundations where people from different domains and groups of society meet to discuss ethical questions concerning scientific and technology issues. Computer scientists are mainly oriented towards getting technical results. It’s an issue we should try to approach together. Take the example of a self-driving car. Who’s responsible for its choices when it’s driving? We need to find those answers together. I think there’s great responsibility on people from the humanities – for example sociologists or philosophers - because they have the conceptual means which people from the technological sectors often lack. The ones who study human nature should also contribute to the creation of the right environment for this discussion.

Do you think that our society is ready to cope with such a challenge like the one represented by robots as substitutes for human beings?
It is likely these technologies would be developed anyway and it is certainly better for society to discuss these questions rather than be surprised when they finally become part of life. Basically, I think we’re developing general artificial intelligence because we don’t understand our own, as well as its relation to consciousness, so we want conscious and intelligent computers to solve this problem for us.

(Revision: Jemma Prior)


Related Articles

Dalla California alla Val di Fassa per documentare il ladino

Un gruppo di studenti della University of Southern California, seguiti dal ricercatore Alessandro Vietti, sta applicando innovativi strumenti computazionali allo studio delle lingue minoritarie e in via di estinzione.

Was darf Forschung mit Social Media Daten?

Wer Daten aus sozialen Medien verwenden darf und für welchen Zweck, wird heftig debattiert, seit die Beratungsfirma Cambridge Analytica mithilfe von Facebook-Daten versucht hat, die Wahlentscheidung von Millionen Amerikanern zu manipulieren. Wissenschaftler jedoch nutzen soziale Netzwerke schon lange als Datenquelle. Auch der Computerlinguist Egon Stemle und der Ökosystemforscher Lukas Egarter Vigl von Eurac Research. Ein Gespräch über Möglichkeiten und Methoden solcher Forschung – und über die ethische Verantwortung des Forschers.

Südtirols Nachwuchsforscher erfolgreich

Mit einem Astronomie- und einem Robotikprojekt überzeugten die Südtiroler Nachwuchsforscher die Jury des Wettbewerbs, an dem Schüler aus Südtirol, dem Trentino, Tirol und Graubünden beteiligt waren. Das Finale fand am 19. und 20. April an der Universität Innsbruck statt. Platz eins in der Kategorie „Forschung“ belegten Schüler des Realgymnasiums Bruneck für die Entdeckung einer bisher nicht bekannten Eigenschaft eines Sterns. Der zweite Platz in der Kategorie „Produktentwicklung“ ging an eine fünfköpfige Schülergruppe vom wissenschaftlichen Gymnasium Rainerum für die Entwicklung des Roboters HAIDI. Die Sieger freuen sich über Geldpreise bis zu 3.000 Euro, zur Verfügung gestellt von der Stiftung Südtiroler Sparkasse.

Infos im Internet: Was ist wichtig, was ist richtig?

Das Internet sammelt das Wissen der Welt, Informationen stehen in Überfülle bereit. Nur: Bekommen wir im Netz auch die richtigen Antworten auf unsere Fragen, enthalten sie alle relevanten Informationen? Dieses Problems nimmt sich die Fakultät für Informatik der Freien Universität Bozen an und hat auch schon interessante Lösungswege gefunden.