emotion perception and recognition: an exploration of cultural differences and ... · 2018. 6....

25
Emotion Perception and Recognition: An Exploration of Cultural Differences and Similarities Vladimir Kurbalija Department of Mathematics and Informatics, Faculty of Sciences, University of Novi Sad Trg Dositeja Obradovića 4, 21000 Novi Sad, Serbia +381 21 4852877, [email protected] Mirjana Ivanović Department of Mathematics and Informatics, Faculty of Sciences, University of Novi Sad Trg Dositeja Obradovića 4, 21000 Novi Sad, Serbia +381 21 4852877, [email protected] Miloš Radovanović Department of Mathematics and Informatics, Faculty of Sciences, University of Novi Sad Trg Dositeja Obradovića 4, 21000 Novi Sad, Serbia +381 21 4852877, [email protected] Zoltan Geler Department of Media Studies, Faculty of Philosophy, University of Novi Sad dr Zorana Đinđića 2, 21000 Novi Sad, Serbia +381 21 4853918, [email protected] Weihui Dai School of Management, Fudan University Shanghai 200433, China [email protected] Weidong Zhao School of Software, Fudan University Shanghai 200433, China [email protected] Corresponding author: Vladimir Kurbalija, tel. +381 64 1810104

Upload: others

Post on 23-Mar-2021

10 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Emotion Perception and Recognition: An Exploration of Cultural Differences and ... · 2018. 6. 27. · Emotion recognition is one of the most challenging tasks ... examined, demonstrating

Emotion Perception and Recognition: An Exploration of Cultural Differences and Similarities

Vladimir Kurbalija

Department of Mathematics and Informatics, Faculty of Sciences, University of Novi Sad

Trg Dositeja Obradovića 4, 21000 Novi Sad, Serbia

+381 21 4852877, [email protected]

Mirjana Ivanović

Department of Mathematics and Informatics, Faculty of Sciences, University of Novi Sad

Trg Dositeja Obradovića 4, 21000 Novi Sad, Serbia

+381 21 4852877, [email protected]

Miloš Radovanović

Department of Mathematics and Informatics, Faculty of Sciences, University of Novi Sad

Trg Dositeja Obradovića 4, 21000 Novi Sad, Serbia

+381 21 4852877, [email protected]

Zoltan Geler

Department of Media Studies, Faculty of Philosophy, University of Novi Sad

dr Zorana Đinđića 2, 21000 Novi Sad, Serbia

+381 21 4853918, [email protected]

Weihui Dai

School of Management, Fudan University

Shanghai 200433, China

[email protected]

Weidong Zhao

School of Software, Fudan University

Shanghai 200433, China

[email protected]

Corresponding author: Vladimir Kurbalija, tel. +381 64 1810104

Page 2: Emotion Perception and Recognition: An Exploration of Cultural Differences and ... · 2018. 6. 27. · Emotion recognition is one of the most challenging tasks ... examined, demonstrating

ABSTRACT

The electroencephalogram (EEG) is a powerful method for investigation of different cognitive processes.

Recently, EEG analysis became very popular and important, with classification of these signals standing

out as one of the mostly used methodologies. Emotion recognition is one of the most challenging tasks

in EEG analysis since not much is known about the representation of different emotions in EEG signals.

In addition, inducing of desired emotion is by itself difficult, since various individuals react differently to

external stimuli (audio, video, etc.). In this article, we explore the task of emotion recognition from EEG

signals using distance-based time-series classification techniques, involving different individuals

exposed to audio stimuli. Furthermore, since some of the participants in the experiment do not

understand the language of the stimuli, we also investigate the impact of language understanding on

emotion perception. Using time-series distances as features for the construction of new data

representations, applied here for the first time to emotion recognition and related tasks, lead to

excellent classification performance, indicating that differences between EEG signals can be used to

build successful models for recognition of emotions, individuals, and other related tasks. In the process,

we observed that cultural differences between the subjects did not have a significant impact on the

recognition tasks and models.

KEYWORDS

Time-series analysis, Classification, Electroencephalogram, Emotion recognition, Cultural differences

1. INTRODUCTION

Emotions represent a multi-disciplinary topic traditionally studied by philosophy, psychology, sociology,

neuroscience, medicine, etc. Due to recent developments in the new research direction of social and

affective computing (Tao, 2005), (Picard, 1997), emotions are now treated as an inter-disciplinary area,

cross-cutting multiple subjects.

Human emotions are considered to be caused by specific situations and activities. Emotional change can

form a unique subjective experience and trigger a series of physiological responses through the nervous

system. By detecting different physiological signals and external features, it is possible to compute and

identify the emotional state of a subject to a certain extent.

Recently, emotion recognition and simulation have become an important research topic in man-

machine communication. Emotional intelligence deals with modeling, recognition, and control of human

emotions. It provides the key theory and essential basis for further studies of a new generation of

intelligent information processing, intelligent services, and similar applications. One of the

contemporary research areas based on emotions and corresponding theories is emotion recognition in

human voice. Therefore, it is reasonable to apply speech signal processing and pattern recognition

techniques, such as cepstrum analysis (Childers, 1997), dynamic time warping (Myers, 1981), and hidden

Page 3: Emotion Perception and Recognition: An Exploration of Cultural Differences and ... · 2018. 6. 27. · Emotion recognition is one of the most challenging tasks ... examined, demonstrating

Markov modeling (Newberg, 2009), to perform different tasks in this area. The ability to correctly

interpret emotional signals is extremely important for the development of intelligent systems and their

applications in a wide range of real-life scenarios.

Five essential domains for emotional intelligence were defined in (Salovey, 1990): knowing one’s

emotions, managing emotions, motivating oneself, recognizing others’ emotions, and handling

relationships. Therefore, emotional intelligence deals with one’s ability to perceive, understand,

manage, and express emotions within oneself and in dealing with others. Having in mind the importance

of proper emotion detection, recognition, and employment in different intelligent environments, it is

necessary to intensify research efforts in these areas and perform experiments on numerous

appropriate data sets.

Recently, we conducted our first real-life experiments with several Serbian and Chinese subjects. During

recordings, EEG brain signals were measured as reactions to short vocal sentences in different emotional

states, including happiness, jealousy, anger, and sadness, all pronounced by a native Mandarin speaker.

In this article, we use this data and extend our previous study (Kurbalija, 2015), the aim of which was to

investigate whether language understanding has an impact on emotion perception, and whether people

of different nationalities “feel” the emotions differently, with the end-goal of obtaining conclusions on

cultural differences and similarities in emotion perception and recognition. We augment the rather

negative results reported in (Kurbalija, 2015) with two new experimental phases that use time-series

distances as features for the construction of new data representations, proposed in (Tomašev, 2015),

(Buza, 2015), and applied here for the first time to emotion recognition and related tasks. The superior

performance of the new representations indicates that the previous negative results stemmed from a

straightforward application of similarity measures to signal time-series, and that differences between

EEG signals can be used to build successful models for recognition of emotions, individuals, and other

related tasks. Furthermore, we observed that cultural differences between the subjects did not have a

significant impact on the recognition tasks and models.

The rest of the paper is organized as follows. Related work is presented in Section 2. Section 3 outlines

the data collection and pre-processing methods. Section 4 presents the experimental setup.

Experimental results on emotion recognition and related classification-based tasks are discussed in

Section 5. Finally, Section 6 draws the conclusions based on the presented work.

2. RELATED WORK

Realizing the fact that by influencing our decision making, learning, and interaction, emotions play an

important role in our everyday life, affective computing, a branch of artificial intelligence, endeavors to

empower machines with the ability to detect, recognize, interpret, and express emotions (Picard, 1997).

Three main approaches to automatic emotion recognition can be distinguished (Wang, 2014): (1)

analyzing facial expressions or speech, (2) inspecting peripheral physiological signals (electrocardiogram,

skin conductance, respiration, pulse), and (3) examining brain signals captured from the central nervous

system (electroencephalograph, electrocorticography and functional magnetic resonance imaging). With

Page 4: Emotion Perception and Recognition: An Exploration of Cultural Differences and ... · 2018. 6. 27. · Emotion recognition is one of the most challenging tasks ... examined, demonstrating

the development of dry electrodes ((Chi, 2012), (Huang, 2015)) and portable headsets, investigating

emotion detection based on EEG signals becomes more and more accessible to researchers.

It can be said that emotion recognition from EEG signals is in its relative infancy, although significant

breakthroughs have been made over the years. In an early review (Coan, 2004), the asymmetry in EEG

signals acquired from frontal lobes is discussed, arguing that it may serve as both a moderator and a

mediator of emotion- and motivation-related constructs. Despite the lack of a definitive conclusion on

this subject, the discussion suggests that frontal lobes carry potentially useful information with respect

to emotion recognition, which was confirmed by subsequent works (Baumgartner, 2006), (Horlings,

2008), (Nie, 2011), (Oude, 2006).

At first, research on emotion recognition from EEG signals focused on stimuli from one source, mostly

visual, but sometimes also on audio stimuli (Baumgartner, 2006). Roughly at the same time, two papers

appeared that examined both types of stimuli, as well as their combination (Baumgartner, 2006), (Oude,

2006). In (Baumgartner, 2006), the influence of visual and musical stimuli on brain processing is

examined, demonstrating that music can markedly enhance the emotional experience evoked by

affective pictures. Research in (Oude, 2006) addresses five questions: (1) Is the modality of the stimulus

recognizable from the recorded brain signals? (2) Are emotions recognizable from an EEG? (3) What is

the influence of the modality of the stimulus on the recognition rate? (4) When using only five

electrodes, what would be a good montage for emotion recognition? and (5) What features are

interesting for emotion recognition? The paper provides evidence towards positive answers for

questions 1 and 2, suggesting more difficulty regarding visual stimuli, and providing advice for electrode

montage and feature use involving F3, F4 and Fpz channels. More recently, movie watching was used as

combined stimuli for detection of positive and negative emotions in (Nie, 2011).

Valence and arousal dimensions of EEG signals for emotion classification are explored in detail in

(Horlings, 2008). The dimensions do not provide satisfactory features for the task on their own.

However, a combined representation using extreme values on both dimensions produced much

improved results.

Features can be extracted from time series obtained from EEG signals in various ways. A comprehensive

overview and systematic analysis of different feature extraction methods using machine learning

techniques for feature selection is given in (Jenke, 2014). Successful approaches for the task of emotion

recognition include the discrete wavelet transform (Murugappan, 2010), fractal dimension (Liu, 2010),

and crossings features (Petrantonakis, 2010a), with the Support Vector Machines (SVM) being among

the most successful classifiers (Wang, 2014).

In (Iacoviello, 2015), an automatic real-time classification method of EEG signals from self-induced

emotions was proposed. The EEG data was collected from ten males between 25 and 40 years old. The

stimuli were generated through disgust produced by remembering unpleasant odor. In the pre-

processing phase, noise was filtered out by using the wavelet transform. After feature selection by

Principal Component Analysis (PCA), the signals were classified applying SVMs, yielding average accuracy

greater than 90%.

Page 5: Emotion Perception and Recognition: An Exploration of Cultural Differences and ... · 2018. 6. 27. · Emotion recognition is one of the most challenging tasks ... examined, demonstrating

In (Yoon, 2013), Yoon and Chung focused on the problem of recognizing human emotions in the valence

and arousal dimensions from EEG signals. They used Fast Fourier Transform (FFT) analysis for feature

extraction and the Pearson correlation coefficient for feature selection. They applied a probabilistic

classifier based on Bayes' theorem and supervised learning using a perceptron convergence algorithm.

In both valence and arousal dimensions, the emotions were defined as two-level classes (high and low)

and three-level classes (high, medium, low). In the two-level class case, the proposed method gave an

average accuracy of 70.9% (valence) and 70.1% (arousal). In the three-level class case, the average

accuracy was 55.4% (valance) and 55.2% (arousal). These examinations were based on the DEAP

database presented by Koelstra et al. (Koelstra, 2012) which includes the EEG and peripheral

physiological signals of 32 participants watching music video clips.

Considering that most of the EEG emotion recognition studies encompassed adults, (Othman, 2013)

analyzed EEG data collected from 5 preschoolers aged 5, watching emotional faces from the Radboud

Faces Database (RafD) (Langner, 2010). These investigations included two-dimensional models of

emotions: recalibrated Speech Affective Space Model (rSASM) (Scherer, 2005) and 12-Point Affective

Circumplex (12-PAC) (Russell, 1980). Features were extracted using Kernel Density Estimation (KDE) and

Mel-Frequency Cepstral Coefficients (MFCC) and classified with a Multi-Layer Perceptron (MLP). The

obtained results supported the 12-PAC model as the preferable one for both feature extraction

methods.

Hybrid Adaptive Filtering (HAF), a novel filtering procedure for efficient isolation of emotion-related

EEG-characteristics was presented in (Petrantonakis, 2010b). HAF was developed by applying genetic

algorithms to the representation of EEG signals based on Empirical Mode Decomposition (EMD) (Huang,

1998) and it was combined with Higher Order Crossings (HOC) (Petrantonakis, 2010b). The classification

efficiency of the HAL-HOC scheme was tested using EEG data retrieved from 16 volunteers in the age

group of 19–32 years while watching pictures with emotion-related facial expressions from the Pictures

of Facial Affect (POFA) database. Six basic emotions were covered by these images: happiness, surprise,

anger, fear, disgust, and sadness. The HAL-HOC scheme outperformed several other approaches

providing classification accuracies of up to 85.17%.

Jenke et al. (Jenke, 2014) investigated three kind of features (power spectrum, wavelet, and nonlinear

dynamic analysis) to assess the association between EEG data and emotional states of six 18–25 year old

subjects (3 male and 3 female) relying on a movie induction experiment that spontaneously leads

subjects to real emotional states. In order to remove noise from the EEG signals, they introduced a novel

linear dynamic system (LDS) based feature smoothing method. Their conclusion is that the power

spectrum feature is superior to other two kinds of features and that the LDS-based smoothing technique

can significantly improve classification accuracy. Furthermore, the obtained results indicated that high

frequency bands play a more important role in emotion activities than low frequency bands.

An investigation of critical frequency bands and channels for efficient EEG-based emotion recognition

utilizing Deep Belief Networks (DBN) is presented in (Zheng, 2015). Three categories of emotions

(positive, neutral, and negative) were covered in this research. EEG signals were recorded from 15

subjects while they were watching emotional movies. The obtained results showed that the proposed

Page 6: Emotion Perception and Recognition: An Exploration of Cultural Differences and ... · 2018. 6. 27. · Emotion recognition is one of the most challenging tasks ... examined, demonstrating

DBN-based approach can produce higher classification accuracy and lower standard deviation than

other feature-based shallow models like kNN, logistic regression, and SVM.

Daly et al. (Daly, 2015) proposed a combination of EEG measures and acoustic properties of music to

predict the emotional response in listeners. 31 individuals (13 male and 18 female) between the ages of

18 and 66 participated in this study. The EEG signals were recorded from 19 channels positioned

according to the International 10-20 system (Jurcak, 2007) for electrode placement. The outcome of the

experiments affirmed that the combined approach performs better in this task than using either of

these feature types alone. In addition, valence was observed to be best-predicted by EEG features in the

delta (0-4Hz) and beta (13-30Hz) frequency bands.

With the aim of supporting the development of a noninvasive assessment tool for automatic detection

of musical emotions, Shahabi and Moghimi (Shahabi, 2016) studied the brain regions associated with

joyful, melancholic, and neutral music through connectivity analysis. Using multivariate autoregressive

modeling, they extracted connectivity patterns among EEG electrodes in different frequency bands. EEG

recordings were performed with 19 nonmusician participants (8 male, 11 female) with similar

educational backgrounds and of age between 18 and 25. They observed that, under the influence of

joyful music, connectivity was increased in the frontal and frontal-parietal regions.

In the last years, another research field within affective computing is gaining more and more attention:

the development of emotional, intelligent software agents. An overview of the state-of-the-art in this

domain is given in (Ivanovic, 2015) along with several possible real-life applications and a discussion

about the required research activities for designing an agent-based architecture, in which agents are

capable of reasoning about and displaying emotions.

The main characteristics of similar works are summarized in the Table 1. The most obvious fact is that all

studies suffer from a relatively small number of participants. That deficiency is also present in our work

especially as we performed experiments with participants from 2 countries from different continents.

This fact suggests that EEG equipment is still scarce and not widely available to the research community,

thus hindering frequent use and larger numbers of participants in the experiments.

All emotion related researches need some kind of stimuli to induce particular participants’ emotion.

Most of the existing studies are rely on audio (music) and video (pictures and movies) stimuli, with one

exception where sense of smell is used. Our research also used audio stimuli. Unlike similar researches

where (international) music is used, the stimuli in our research are vocal sentences presented in the

Mandarin language. One group of participants consists of native Mandarin speakers, while the other

group does not understand the presented language. With this experimental set-up, we wanted to

investigate whether language understanding have the impact on induced emotions.

Page 7: Emotion Perception and Recognition: An Exploration of Cultural Differences and ... · 2018. 6. 27. · Emotion recognition is one of the most challenging tasks ... examined, demonstrating

Paper Participants Stimuli Techniques

A real-time classification algorithm for EEG-based BCI driven by self-induced emotions (Iacoviello, 2015)

10 Unpleasant odor

Wavelets, PCA, SVM

EEG-based emotion estimation using Bayesian weighted-log-posterior function and perceptron convergence algorithm (Yoon, 2013)

32

Music video clips

FFT, Pearson correlation, Bayes' classifier, perceptron

EEG Emotion Recognition Based on the Dimensional Models of Emotions (Othman,2013) 5

Pictures of emotional faces

KDE, MFCC, MLP

Emotion Recognition from Brain Signals Using Hybrid Adaptive Filtering and Higher Order Crossings Analysis (Petrantonakis, 2010b)

16 Pictures of emotional faces

Genetic algorithms, EMD, HOC, HAF

Feature Extraction and Selection for Emotion Recognition from EEG (Jenke, 2014)

6 Emotional movie

Power spectrum, wavelets, NDA, LDS

Investigating Critical Frequency Bands and Channels for EEG-Based Emotion Recognition with Deep Neural Networks (Zheng, 2015)

15 Emotional movie

DBN, kNN, logistic regression, SVM

Music-induced emotions can be predicted from a combination of brain activity and acoustic features (Daly, 2015)

31 Music Acoustic descriptors

of the music, regression

Toward automatic detection of brain responses to emotional music through analysis of EEG effective connectivity (Shahabi, 2016)

19 Music multivariate

autoregressive modeling

Table 1. Characteristics of similar works

All related researches firstly perform different kinds of feature extraction (PCA, FFT, KDE etc.) and

afterwards classification or similar data mining tasks for evaluation of the model. Our approach is

different: we use original time series received from EEG equipment and perform 1NN classification with

different similarity measures. In such approach, we try to identify similar behavior of brain waves in

terms of similar time series trends or patterns.

3. DATA PREPROCESSING

Although emotion recognition from EEG signals is in its early stage of understanding, some techniques

(especially those related to data preprocessing) already became standard. Instead of analyzing EEG

signals directly, the usual approach in emotion detection algorithms is to decompose signals into distinct

frequency bands, and then work on each band separately. Although the exact frequency ranges vary in

different sources, the following set is most commonly used (Mikhail, 2010)(Murugappan, 2010): theta

(4-8 Hz), alpha (8-14 Hz), beta (14-32 Hz), and gamma (32-64 Hz). Lower frequencies (delta, 1-4 Hz) are

associated with babies and sleeping adults, while higher frequencies (i.e. above 64 Hz) represent noise.

Frequency bands can be extracted from raw EEG signals using band-pass audio filters (Shenoi, 2005).

These filters pass all frequencies within the given range, while discarding frequencies that fall out of that

range. In practice, the filter is configured through the central frequency and the band width parameters.

Page 8: Emotion Perception and Recognition: An Exploration of Cultural Differences and ... · 2018. 6. 27. · Emotion recognition is one of the most challenging tasks ... examined, demonstrating

In case of the alpha band, for example, the central frequency would be 11 Hz, while the band width

would be 3 Hz.

In order to pre-process our data and extract the required frequency bands, we used the two-pole

Butterworth filter (Bianchi, 2007) as the concrete implementation of the band-pass filter. This

implementation is freely available in FFmpeg, a collection of libraries and tools for multimedia

processing (FFmpeg, 2015). Figure 1 shows the results of filtering a signal collected for one of the

experiment subjects. The bottom-most sub-figure shows the raw EEG signal, while the middle and the

top sub-figures show the theta and alpha frequency bands, respectively, that were obtained through

data filtering.

Figure 1. A part of a original EEG signal (bottom) and the corresponding theta (middle) and alpha (top)

frequency bands obtained by using the two-pole Butterworth band-pass filter

4. EXPERIMENTAL SETUP

In order to perform the experiments and discover if there are differences in emotion recognition

between people from different socio-cultural settings, we acquired the data using a rather standard

procedure. During the data acquisition process, there were six participants: four native Mandarin

speakers and two native Serbian speakers. Scalp electrodes were applied using the 10-20 system, which

is one of the most widely-used international standards (Jurcak, 2007). It typically includes 21 electrodes,

with distances between adjacent electrodes representing 10% or 20% of the total front-to-back or left-

to-right surfaces. Each electrode placement site has a letter to identify the lobe, or area of the brain it is

reading from: Pre-frontal (Fp), Frontal (F), Temporal (T), Parietal (P), Occipital (O), and, Central (C). Even

numbers refer to electrode placement on the right side of the head, odd numbers refer to those on the

left side, whereas a "Z" (zero) refers to an electrode placed on the midline. The reference electrodes

were fixed at the subject’s auricle (A1 and A2 channels).The positions of electrodes are depicted

graphically in Figure 2, with the detail alphanumeric nomenclature being described in (Society, 2006).

Page 9: Emotion Perception and Recognition: An Exploration of Cultural Differences and ... · 2018. 6. 27. · Emotion recognition is one of the most challenging tasks ... examined, demonstrating

Figure 1. Electrodes of the international 10-20 standard

Instead of acquiring the data from electrodes directly, it is common to use the so-called bipolar model,

which measures the difference between a pair of electrodes (Russell, 1979). Therefore, to perform

experiments with members of our two groups, we constructed standard 8 channels, marked C1 to C8, as

follows: C1: Fp1-T3, C2: Fp2-T4, C3: T3-O1, C4: T4-O2, C5: Fp1-C3, C6: Fp2-C4, C7: C3-O1, C8: C4-O2.

4.1. Extraction of time series

Experiments for data acquisition were performed in a relatively controlled environment, using the

described electrodes. The main idea was that Chinese and Serbian participants listen to audio clips

(mainly the same sentences pronounced on Mandarin language) which express different kinds of

emotions, while their brain activity during experiments was recorded using 8 channels (marked as C1 to

C8) of combined electrodes fastened to their scalp. The audio clips were uniformly structured as

presented in Table 2. All participants were right-handed and without serious eyesight problems.

Event Number

Stimulating Signal Language Description

1 Speech (anger) Mandarin

2 Speech (fear) Mandarin

3 Speech (happiness) Mandarin

4 Speech (neutral) Mandarin

5 Speech (sadness) Mandarin

6 Speech (surprise) Mandarin

7 Music (Chinese classic song)

Mandarin The joy of love

8 Music (jazz) English Fly me to the moon

9 Music (Chinese traditional instruments)

Mandarin Butterfly Lovers

10 Music (rock and roll) English Breed

Table 2. The specification of audio clips

Page 10: Emotion Perception and Recognition: An Exploration of Cultural Differences and ... · 2018. 6. 27. · Emotion recognition is one of the most challenging tasks ... examined, demonstrating

The first six events represent several sentences performed by a Chinese actor simulating different

emotions: angry, fear, happy, neutral, sad, and surprise. The same set of sentences is reproduced with

appropriate intonation in all six emotions. The purpose of these events is to discover whether the way of

pronunciation affects the perception of emotions regardless of speech understanding. The last four

events represent different music clips (traditional, jazz, rock) whose purpose was to investigate whether

the same music causes the same brain activity, or emotion, in different participants.

During the reproduction of audio clips, brain activity was recorded using 8 channels (C1 to C8). The part

of recorded data for one participant is given in Table 3. Sample rate for this experiment was set to

125Hz. All 10 events were presented to each of six participants and brain reactions were recorded. The

number of samples/rows per participant varies from 21810 to 102914. The reason for this difference is

that participants were not presented the same subset of sentences from a certain group/emotion.

Event

Number Min. Sec. C1 C2 C3 C4 C5 C6 C7 C8

1 1.241 74.43 -2.254 -25.359 20.626 -17.921 -3.381 -17.244 13.187 -25.359

1 1.241 74.438 20.4 -14.539 27.726 -32.122 11.158 -25.359 10.482 -26.712

1 1.241 74.445 1.916 -36.855 7.777 -40.575 26.036 -26.712 24.683 -38.208

1 1.241 74.453 34.601 -9.129 41.927 -11.496 63.229 -6.424 33.474 -20.626

1 1.241 74.461 19.048 -29.755 10.144 -42.265 22.654 -19.611 19.949 -16.568

1 1.241 74.469 16.793 -4.057 23.331 7.101 21.978 11.158 31.445 -5.072

1 1.241 74.477 12.511 -.338 -2.367 -12.511 -10.144 -17.582 5.41 -8.453

1 1.241 74.484 -19.724 -22.654 -22.992 -18.259 -3.719 -18.935 12.849 2.029

1 1.242 74.492 14.765 3.719 30.769 10.144 32.122 1.014 23.331 -1.352

1 1.242 74.5 -.902 -37.87 2.029 -42.942 3.043 -39.56 25.021 -18.597

Table 3. The structure of raw data for one participant

As discussed earlier, the collected data in the raw form is not suitable for direct analysis. The EEG signals

from each channel are firstly decomposed to its alpha, beta, gamma, and theta frequency band using

the two-pole Butterworth filter. To obtain a meaningful form of time series appropriate for further

analysis, each signal from each channel and each frequency band is split into 10 time series on the basis

of the performed event. Finally, we obtained a labeled dataset which consists of 1920 time series. Time

series are labeled with different labels for the purpose of further experiments including:

Event number (emotion) from Table 2.

Name of the participant.

Nationality of the participant.

Channel (C1 to C8).

Frequency band (alpha, beta, gamma, and theta).

Page 11: Emotion Perception and Recognition: An Exploration of Cultural Differences and ... · 2018. 6. 27. · Emotion recognition is one of the most challenging tasks ... examined, demonstrating

5. EXPERIMENTS

With the described experimental setup, we performed several sets of experiments, all with the purpose

of discovering similarities/differences in emotion perception in different language participants.

Figure 3. Workflow of experimental phases

The broad structure of all experiment phases is given in Figure 3. The raw data from EEG headset is first

filtered using FFmpeg software with two-pole Butterworth filter. Filtered data is labeled according to

previously described methodology. Labeled time series dataset prepared in such a way is processed one

or two times using FAP (Framework for Analysis and Prediction) system (Kurbalija, 2010). The results

from phases 1 and 2 are obtained after one utilization of FAP system using: 10 runs of 10-fold Stratified

Cross Validation (SCV10x10), 1 Nearest Neighbor classifier (1NN) with Dynamic Time Warping (DTW)

distance measure. For phases 3 and 4, the first utilization of FAP produced a newly formed labeled time-

series datasets: emotion-difference time series dataset and participant-difference time series dataset.

These datasets are processed one more time with FAP system in order to obtain final results. For phases

3 and 4, two more distance measures were used: Euclidean distance (L2) and Edit Distance with Real

Penalty (ERP); while the rest of experimental setup is the same. All phases will be described in detail in

the following subsections.

5.1. Phase 1

Within the first phase, we performed two sets of experiments. The experimental setup was the same for

both sets, while only the used labels of time series were different.

In the first set of the investigations, the time series were labeled relying solely on the emotion (the

event number) regardless of the participants. The aim of these experiments was to examine whether

there are similarities or common patterns between the time series of different participants for the same

emotion. This could indicate that different people experience the same emotion in similar way,

independently of their nationality or language understanding.

In the second phase of the analysis, we have incorporated the information about the nationality of the

participants into the class labels, too. The motivation for this extension was in the assumption that there

Page 12: Emotion Perception and Recognition: An Exploration of Cultural Differences and ... · 2018. 6. 27. · Emotion recognition is one of the most challenging tasks ... examined, demonstrating

may be significant differences in the responses of the participants of different nationalities which lead to

considerably dissimilar time series.

The common methodology for estimating the similarities between time series is to measure

classification accuracy of labeled time series (Ding, 2008), (Kurbalija, 2014). For the evaluation

technique, we used 10 runs of 10-fold Stratified Cross Validation (SCV10x10). The simple 1NN classifier

was used since it was shown that it gives among the best results (compared to many not only distance-

based classifiers) with time-series data (Ding, 2008). As the distance measure for this distance-based

classifier, the Dynamic Time Warping (DTW) (Berndt, 1994) measure was chosen as one of the most

prominent ones in time-series mining. Another reason for choosing DTW is its natural fit to the scenario

where time series are of different length. The time series were z-normalized before the experiments. All

the experiments were performed using FAP system (Kurbalija, 2010).

In the first set of experiments, we measure classification error rate on the prepared time-series dataset,

where the labels of time series consist only of events (6 emotions and 4 kinds of music). The original

dataset is divided into 32 sub-datasets, where each sub-dataset contains time series with the same

channel and frequency band. Each sub-dataset contains 60 time series: 6 participants with 10 events. In

each sub-dataset, SCV10x10 is performed with a 10-class classifier. The purpose of this experiment was

to eventually determine which combination of channel and frequency band gives the best results in

emotion recognition. The good results would be indicated by relatively small classification error rate.

The results for all 32 sub-datasets are given in Table 4.

Channel Wave Err. Rate Channel Wave Err. Rate

C3 theta 0.871667 C4 alpha 0.923333

C3 alpha 0.876667 C4 gamma 0.923333

C4 beta 0.878333 C1 gamma 0.925

C5 theta 0.878333 C5 gamma 0.926667

C8 alpha 0.88 C6 gamma 0.926667

C3 gamma 0.893333 C2 gamma 0.928333

C8 theta 0.893333 C7 gamma 0.928333

C6 beta 0.896667 C1 alpha 0.93

C2 alpha 0.905 C1 theta 0.938333

C4 theta 0.908333 C6 alpha 0.938333

C7 alpha 0.908333 C8 gamma 0.94

C2 theta 0.911667 C5 beta 0.945

C6 theta 0.911667 C7 beta 0.956667

C8 beta 0.913333 C2 beta 0.958333

C5 alpha 0.918333 C1 beta 0.963333

C7 theta 0.921667 C3 beta 0.975

Table 4. The error rates of classification with respect to emotion

Page 13: Emotion Perception and Recognition: An Exploration of Cultural Differences and ... · 2018. 6. 27. · Emotion recognition is one of the most challenging tasks ... examined, demonstrating

The error rate of a 10-class random classifier with equal distribution of classes is 0.9. The presented

results show that all the combinations are in the level of a random classifier, and not one combination of

channel and frequency band stands out. This means that the most similar time series from the training

set (according to DTW) most often does not belong to the same class as the time series whose class we

are trying to predict. In other words, none of these combinations can be used to effectively distinguish

between recordings of the participants' brain activity induced by listening to audio clips which express

different kinds of emotions.

In the second set of experiments, the division into 32 sub-datasets is the same as in the first setup. The

only difference is that the classification is performed according to events and nationality of participant.

This represents 20-class classification (2 nationalities: Chinese and Serbian with 10 events). In this setup,

the labels are not equally distributed: each sub-dataset contains recordings of four Chinese and two

Serbian subjects.

By this evaluation, we intended to investigate whether language understanding has an impact on

emotion perception and whether different nationalities “feel” the emotions differently. The results are

given in Table 5.

Channel Wave Err. Rate Channel Wave Err. Rate

C3 alpha 0.89 C6 theta 0.946667

C3 theta 0.895 C2 beta 0.961667

C5 theta 0.896667 C3 gamma 0.961667

C6 beta 0.896667 C7 beta 0.961667

C4 beta 0.911667 C4 alpha 0.963333

C7 theta 0.923333 C7 alpha 0.966667

C4 theta 0.926667 C3 beta 0.971667

C8 theta 0.926667 C1 beta 0.976667

C8 beta 0.928333 C1 theta 0.976667

C2 theta 0.93 C2 gamma 0.978333

C1 alpha 0.931667 C6 gamma 0.978333

C8 alpha 0.931667 C4 gamma 0.981667

C5 beta 0.94 C1 gamma 0.993333

C5 alpha 0.941667 C5 gamma 0.995

C6 alpha 0.941667 C7 gamma 0.995

C2 alpha 0.946667 C8 gamma 0.998333

Table 5. The error rates of classification with respect to emotion and nationality

The results are again at the level of a random classifier (for a 20-class random classifier with equal

distribution of classes the error rate is 0.95) with no specific emphasized combination. This means that

even in this case no specific combination (of wave and channel) can be successfully used to differentiate

time series corresponding to distinct classes. The obtained classification accuracies indicate that the

Page 14: Emotion Perception and Recognition: An Exploration of Cultural Differences and ... · 2018. 6. 27. · Emotion recognition is one of the most challenging tasks ... examined, demonstrating

brain signals recorded under the influence of listening to audio tracks of different emotions are not

considerably different, according to the DTW measure taken on raw signals.

5.2. Phase 2

In the next experimental phase, we wanted to test whether there are differences between brain signals

of participants with regard to their nationality. The emotions were not taken into consideration. The

experimental setup was very similar to the previous one:

The dataset is divided into 32 sub-datasets, where each sub-dataset contains time series with

the same channel and frequency band. Each sub-dataset contains 60 time series: 6 participants

with 10 events.

1NN classifier was used.

DTW measure was used for computing distances between time series.

SCV10x10 was used as the evaluation technique.

The only difference here is that we used a 2-class classifier – only nationalities of participants (Serbian

and Chinese) are considered as labels for 1NN classifier. The results for all 32 sub-datasets, sorted by

error rates, are shown in Table 6.

Channel Wave Err. Rate Channel Wave Err. Rate

C8 gamma 0.261667 C5 alpha 0.391667 C7 gamma 0.295 C2 beta 0.391667

C1 gamma 0.306667 C8 beta 0.396667

C6 gamma 0.308333 C7 beta 0.398333

C5 gamma 0.31 C1 beta 0.406667

C2 gamma 0.31 C4 beta 0.411667

C4 gamma 0.311667 C4 theta 0.413333

C3 gamma 0.323333 C4 alpha 0.416667

C1 theta 0.353333 C2 theta 0.43

C5 beta 0.371667 C7 alpha 0.435

C8 alpha 0.371667 C2 alpha 0.436667

C7 theta 0.375 C6 beta 0.436667

C5 theta 0.381667 C8 theta 0.436667

C3 alpha 0.385 C6 alpha 0.445

C3 theta 0.386667 C6 theta 0.446667

C3 beta 0.39 C1 alpha 0.451667

Table 6. The results of classification with respect to nationality

The classes in each dataset are not distributed equally, but in proportion 2:1 (Chinese:Serbian).

Therefore, the error rate of a random classifier is 0.33. Again, the results are in the range of a random

classifier. However, one interesting fact is evident from these results: the best 8 results, and only results

better than random classifier, are achieved with gamma waves. This could suggest that there are

significant differences between time series of gamma waves of participants with different nations, or

Page 15: Emotion Perception and Recognition: An Exploration of Cultural Differences and ... · 2018. 6. 27. · Emotion recognition is one of the most challenging tasks ... examined, demonstrating

that the gamma waves are responsible for language understanding, since only one group of participants

(Chinese) can understand the language (Mandarin) of audio signals.

5.3. Phase 3

Following the observations from the previous phases, we limit our investigation only to audio clips

which are connected to 6 emotions (events 1–6). Thus, the audio clips and the corresponding brain

waves associated with music (events 7–10) are completely discarded in this and following phases. The

reason for this is the observation that it is not clear which kind of emotion is induced in particular

subject as a reaction to a particular kind of music.

In addition to that, the length of time series is shortened in the following manner: the shortest time

series in the dataset was identified, and only the last 5 seconds were extracted as a new time series. This

same 5-second interval was extracted from all time series from the dataset. As a result, the new dataset

contains time series of the same length and the data which belongs to the same interval from the

beginning of recording (this is done for each participant and for each emotion). The main reasons for

these cutoffs are:

It is assumed that a certain amount of time is needed to stabilize the brain waves when a new

emotion is presented, with the time possibly differing with between participants. This cutoff

tries to capture the original waves associated to a particular emotion.

Time series are now of the same length which enables the use of more similarity measures.

With these extracted time series, we compared the time series associated to each emotion. This is done

for each participant, each channel, and each wave. Additionally, three time-series similarity measures

were used: DTW, ERP (Edit Distance with Real Penalty) (Chen, 2004) and the standard L2 metric.

However, different similarity measures on original data give results in different ranges. Therefore, these

time series obtained with one similarity measure will almost always be more similar than the time series

obtained with different measures. Consequently, the time series obtained with different similarity

measures will not be comparable. To cope with this problem, we normalized the obtained time series by

applying min-max normalization on each value of each time series (where minimum and maximum

distances are computed over time series):

Part of the newly obtained dataset is shown in Table 7. The first five columns represent participant,

nationality, similarity measure, channel, and wave, respectively. The remaining columns represent the

similarity between time series of two emotions for a particular participant, channel, and wave. For

example, column E25 contains the similarity between time series of the second and fifth emotion (fear

and sadness) for a given participant and a given channel and wave.

NAME NAT. SIM. CH. WAVE E12 E13 E14 E15 E16 E23 E24 E25 E26 E34 E35 E36 E45 E46 E56

xudong 0 L2 C1 alpha 0.63 0.19 0.61 0.21 0.93 0.78 0.60 0.53 0.70 0.36 0.79 0.80 0.98 0.00 1.00

xudong 0 DTW C1 alpha 1.00 0.67 0.74 0.58 0.49 0.16 0.06 0.01 0.12 0.11 0.01 0.00 0.11 0.13 0.00

Page 16: Emotion Perception and Recognition: An Exploration of Cultural Differences and ... · 2018. 6. 27. · Emotion recognition is one of the most challenging tasks ... examined, demonstrating

xudong 0 ERP C1 alpha 1.00 0.88 0.91 0.72 0.61 0.07 0.03 0.04 0.18 0.14 0.08 0.00 0.10 0.17 0.04

xudong 0 L2 C1 beta 0.60 0.35 0.75 0.44 0.45 0.82 0.89 0.78 0.55 1.00 0.00 0.19 0.53 0.36 0.54

xudong 0 DTW C1 beta 0.91 0.78 1.00 0.95 0.88 0.24 0.23 0.17 0.21 0.06 0.15 0.07 0.03 0.14 0.00

xudong 0 ERP C1 beta 0.84 0.79 1.00 0.80 0.72 0.18 0.05 0.04 0.16 0.04 0.08 0.01 0.08 0.06 0.00

xudong 0 L2 C1 gamma 0.68 0.98 0.82 0.83 0.98 0.73 1.00 0.97 0.47 0.22 0.08 0.15 0.00 0.51 0.47

xudong 0 DTW C1 gamma 0.70 1.00 0.96 0.78 0.90 0.87 0.55 0.00 0.82 0.23 0.54 0.32 0.76 0.28 0.73

xudong 0 ERP C1 gamma 0.80 0.87 1.00 0.76 0.91 0.61 0.52 0.02 0.78 0.34 0.76 0.48 0.93 0.00 0.90

xudong 0 L2 C1 theta 1.00 0.82 0.84 0.98 0.90 0.81 0.79 0.91 0.53 0.23 0.00 0.57 0.48 0.36 0.76

Table 7. Part of the dataset for phase 3

In this newly prepared dataset, we consider data from columns E12 to E56 as features that describe the

given data instances, and since we will be applying time-series distance measures, we will effectively be

treating the data in this new representation as time series of length 15. We will refer to them as

emotion-difference time series. This approach to feature construction was first proposed in (Tomašev,

2015), and applied to time series in (Buza, 2015) to the task of brain disease recognition from EEG data.

To the best of our knowledge, this is the first time this approach has been applied to emotion

recognition.

With emotion-difference time series, we performed testing of classification accuracies, where the

classification is done according to: name, nationality, similarity measure between original time series,

channel, and wave. For these five classification tasks, the standard experimental setup which includes

the SCV10x10 cross validation technique with the 1NN classifier was used. Furthermore, three time-

series distance measures were used: DTW, ERP and L2. The error rates of performed classification testing

tasks are shown in Figure 4 (the number in parenthesis on the x-axes represents the number of classes

for a specified classification task). We performed these experiments with and without applying z-

normalization on the original time series. Min-max normalization of emotion-difference time series was

done in both cases.

a.

3.8% 2.4%

21.9%

85.8%

15.6%

7.9% 2.9%

21.5%

87.9%

19.3%

6.8% 3.3%

22.8%

87.2%

16.2%

0.0%

10.0%

20.0%

30.0%

40.0%

50.0%

60.0%

70.0%

80.0%

90.0%

100.0%

NAME (6) NATIONALITY (2) SIMILARITY (3) CHANNEL (8) WAVE (4)

Erro

r ra

te

Classifier

L2

DTW

ERP

Page 17: Emotion Perception and Recognition: An Exploration of Cultural Differences and ... · 2018. 6. 27. · Emotion recognition is one of the most challenging tasks ... examined, demonstrating

b.

Figure 4. Error rates of classification with emotion-difference time series (a. without and b. with z-normalization of original time series)

The presented results show that all three similarity measures give similar results, where L2 gives the

lowest classification errors and DTW the highest. Slightly better results are obtained with non-

normalized original time series. The most interesting observation is that very high accuracies are

achieved for classification with regard to participant and nationality. The high nationality-accuracy is

related to (and a consequence of) high participant-accuracy. The high participant-accuracy means that

emotion-difference time series are similar for one participant, regardless of channel, wave, and

similarity measure. Consequently, emotion-difference time series are significantly different between

different participants, where clear separation is present. The similar emotion-difference time series of

one participant mean that the similarities between emotion waves are characteristic for one participant

regardless of channel, wave, and used similarity measure. For example, if value E25 (the similarity

between the second and the fifth emotion) is relatively small for one participant, it tends to be relatively

small for all of his/her combinations of channels, waves, and similarity measures. This small value might

not be characteristic of other participants. This reasoning suggests that every participant has his/her

own “ordering” of emotions, i.e. a distribution of similarities between emotions that tend to be different

amongst different participants.

Solid results are obtained also for wave classification. This means that there is relatively clear separation

between wave time series among all participants with regards to emotion. This behavior is somehow

expected because the wave channels are extracted using band-pass audio filters and these time series

affect different frequency ranges. The accuracy according to similarity measure classification could

suggest that some similarity measures are more appropriate for some emotion differences. Finally, the

accuracy of channel classification is the worst, which indicates that channels cannot be differentiated via

differences between emotion time-series.

12.0% 5.7%

21.3%

89.1%

18.0% 20.4%

12.1%

25.6%

91.4%

29.5%

20.8%

11.2%

24.2%

91.6%

28.9%

0.0%

10.0%

20.0%

30.0%

40.0%

50.0%

60.0%

70.0%

80.0%

90.0%

100.0%

NAME (6) NATIONALITY (2) SIMILARITY (3) CHANNEL (8) WAVE (4)

Erro

r ra

te

Classifier

L2

DTW

ERP

Page 18: Emotion Perception and Recognition: An Exploration of Cultural Differences and ... · 2018. 6. 27. · Emotion recognition is one of the most challenging tasks ... examined, demonstrating

5.4. Phase 4

In view of the results from the previous phase, we wanted to further investigate the connection

between each individual and his/her emotion recognition. In this phase, we used a “dual” approach to

the previous phase. For the purposes of feature construction, we compared the time series associated

with each participant (with 5-second extracted time series, the same as previously described). This was

done for each emotion, each channel, and each wave. Again, three time-series similarity measures were

used: DTW, ERP and L2; and the obtained data were min-max normalized. Part of the newly obtained

dataset is shown in Table 8. The first five columns represent emotion, similarity measure, channel, and

wave, respectively. The remaining columns represent the similarity between time series of two

participants for a particular emotion, channel, and wave. For example, column P14 contains the

similarity between time series of the first and the fourth participant for a given emotion, channel, and

wave.

EMOT. SIM. CH. WAVE P12 P13 P14 P15 P16 P23 P24 P25 P26 P34 P35 P36 P45 P46 P56

angry L2 C1 alpha 0.56 0.49 0.55 0.66 0.45 0.13 0.47 0.82 0.37 0.74 0.21 0.00 0.65 0.60 1.00

angry DTW C1 alpha 0.86 0.90 0.43 0.98 1.00 0.13 0.47 0.13 0.06 0.87 0.18 0.08 0.52 0.66 0.00

angry ERP C1 alpha 0.87 1.00 0.62 0.94 0.97 0.33 0.50 0.15 0.16 0.75 0.26 0.00 0.43 0.54 0.07

angry L2 C1 beta 0.79 0.82 0.87 0.80 0.41 0.00 0.78 0.73 0.61 0.84 0.54 0.73 0.97 1.00 0.96

angry DTW C1 beta 0.61 0.59 0.50 0.54 0.48 0.14 1.00 0.07 0.24 0.91 0.00 0.20 0.97 0.99 0.19

angry ERP C1 beta 0.98 1.00 0.66 0.92 0.82 0.12 0.78 0.00 0.11 0.83 0.01 0.12 0.78 0.83 0.11

angry L2 C1 gamma 0.74 1.00 0.93 0.25 0.12 0.21 0.32 0.93 0.47 0.00 0.92 0.85 0.89 0.82 0.51

angry DTW C1 gamma 0.50 0.45 0.97 0.77 0.71 0.00 0.83 0.82 0.69 0.75 0.82 0.74 0.99 1.00 0.52

angry ERP C1 gamma 0.44 0.37 1.00 0.94 0.86 0.00 0.84 0.92 0.78 0.86 0.92 0.83 0.97 0.95 0.64

angry L2 C1 theta 0.45 0.86 0.89 0.45 0.79 0.34 0.00 1.00 0.53 0.36 0.27 0.75 0.36 0.32 0.99

Table 8. Part of the dataset for phase 4

As in the previous phase, we consider data from columns P12 to P56 as data features, obtaining time

series of length 15. We will denote them as participant-difference time series. With these time series,

we performed testing of classification accuracies, where classification is done according to: emotion,

similarity measure between original time series, channel, and wave. For these four classification tasks,

the same experimental setup with the SCV10x10 cross validation technique and the 1NN classifier was

used. Again, three time-series distance measures were used with the 1NN classifier: DTW, ERP and L2.

The experiments are performed with and without applying z-normalization on the original time series.

The error rates of performed classification tests are shown in Figure 5.

Page 19: Emotion Perception and Recognition: An Exploration of Cultural Differences and ... · 2018. 6. 27. · Emotion recognition is one of the most challenging tasks ... examined, demonstrating

a.

b.

Figure 5. Error rates of classification with participant-difference time series (a. without and b. with z-

normalization of original time series)

Again, the results without z-normalization are slightly better, where L2 measure gives the best results

overall. Here, we are most interested in classification accuracy with regard to emotion. Again, the

classification error of around 15% (without z-normalization) is significantly lower compared to a random

6-class classifier which is around 83%. This high accuracy means that the similarity between participant-

difference time series is high for the same emotion, and that there is relatively clear separation.

Consequently, the participant-difference time series tend not to be similar for different emotions. This

again supports the observation that each person has his/her own unique set of emotion waves

(independent from nationality).

14.5%

2.0%

89.1%

18.0% 14.8%

2.3%

88.9%

18.1% 15.5%

3.1%

88.4%

18.4%

0.0%

10.0%

20.0%

30.0%

40.0%

50.0%

60.0%

70.0%

80.0%

90.0%

100.0%

EMOTION (6) SIMILARITY (3) CHANNEL (8) WAVE (4)

Erro

r ra

te

Classifier

L2

DTW

ERP

12.0%

23.3%

85.0%

19.5% 23.1%

28.1%

88.6%

28.7% 23.3% 25.5%

88.3%

28.6%

0.0%

10.0%

20.0%

30.0%

40.0%

50.0%

60.0%

70.0%

80.0%

90.0%

100.0%

EMOTION (6) SIMILARITY (3) CHANNEL (8) WAVE (4)

Erro

r ra

te

Classifier

L2

DTW

ERP

Page 20: Emotion Perception and Recognition: An Exploration of Cultural Differences and ... · 2018. 6. 27. · Emotion recognition is one of the most challenging tasks ... examined, demonstrating

Classification according to wave and channel is similar as in the preceding discussion: wave classification

is better, but not much, while channel classification is the worst. Very high accuracy which is obtained

for similarity-classification (2%-3% for non-normalized original time series) is a direct consequence of

the fact that different similarity measures give results in different ranges. Therefore, the results

obtained with one similarity measure are usually more similar to some other results obtained with the

same similarity measure. Consequently, results obtained with different similarity measures will always

be significantly different. This phenomenon is not present when the original time series are z-normalized

(Figure 5b.), where similarity-classification does not give such unrealistically low error rates.

The presented results clearly show that emotion classification gives the best results (especially with the

L2 metric) when normalized sets are used, especially having in mind that 6-class classification was

performed.

In phases 3 and 4 the time series were of the same length, which enabled the usage of more diverse

time-series similarity measures. Despite the fact that the DTW measure give the most promising results

in time series databases, our results suggest that the simple L2 measure perform slightly better with EEG

data. The computational complexity of DTW and ERP measures is quadratic, while the complexity of L2 is

linear (Kurbalija, 2011). Therefore, our analysis suggests that the L2 measure is convenient for emotion

perception kinds of tasks since it gives qualitatively slightly better results (from the classification

accuracy perspective) and remarkably higher computation performance.

6. CONCLUSION

We presented a complete research cycle involving EEG data acquisition, preprocessing, labeling, and

exploration of different time-series distance-based approaches to signal classification, with the aim of

building successful models for emotion recognition and related tasks.

Data was collected through EEG brain signal recordings involving Serbian and Chinese native speakers,

measuring reactions to short vocal sentences in different emotional states, including happiness,

jealousy, anger, and sadness, all pronounced by a native Mandarin speaker. In this article, we extended

our previous study (Kurbalija, 2015), augmenting the reported negative results with two new

experimental phases that use time-series distances as features for the construction of new data

representations (Tomašev, 2015), (Buza, 2015), applied here for the first time to emotion recognition

and related tasks. The superior performance of the new representations indicated that differences

between EEG signals could be used to build successful models for recognition of emotions, individuals,

and other related tasks. During the whole process, we observed that cultural differences between the

subjects did not have a significant impact on the recognition tasks and models.

The results of our research suggest that there are no significant similarities between EEG signals of

different participants for the same emotion, which is induced by an audio stimulus. It is likely to say that

each individual has its own ordering of emotions which is characteristic of each person. This fact is

supported by high classification accuracies in phases 3 and 4. Additionally, the results of our

Page 21: Emotion Perception and Recognition: An Exploration of Cultural Differences and ... · 2018. 6. 27. · Emotion recognition is one of the most challenging tasks ... examined, demonstrating

experiments suggest L2 measure as the most convenient for emotion related tasks when working with

time series extracted from EEG data. It gives better classification accuracies, while at the same time it

gives significantly lower computational times than other more sophisticated time series measures (DTW

and ERP). However, these conclusions are result of a relatively small case study, while more extensive

experiments are required for some general conclusions.

There are various paths for extending this research in future work. Gathering and using more data,

involving more nationalities and more types of stimuli is one direction, possibly adding more dimensions

of differences between individuals in addition to language/culture. Another direction involves

investigating other classification methodologies, with the goal of improving recognition performance.

The impact on recognition performance should also be investigated by applying non-linear feature

extraction techniques such as: Recurrence Quantification Analysis (RQA), Intrinsic Time-scale

Decomposition (ITD), Higher Order Spectral (HOS) analysis, Empirical Mode Decomposition (EMD), etc.

Finally, combination of the obtained models and insights with other recognition tasks from EEG data, as

well as other types of information recorded from the brain (such as functional magnetic resonance

imaging – fMRI) could help improve the overall understanding of the brain’s functions and abilities,

leading to improvements of existing recognition techniques and possibly the discovery of novel

applications.

ACKNOWLEDGMENTS

This work was partially supported by the Ministry of Education, Science and Technological Development

of the Republic of Serbia, through project no. OI174023: “Intelligent techniques and their integration

into wide-spectrum decision support,” by bilateral project no. 2-9/2013: “Emotional Intelligence and

Applications Based on Multi-Agent Systems,” by bilateral project no. 4-9/2018: “Human­machine

Harmony Emotional Intelligence Based on the Imitative Brain,” and by International Cooperative

Program, Shanghai Municipal Science and Technology Commission, China (No.16550720500).

REFERENCES

Baumgartner, T., Esslen, M. and Jäncke, L. (2006). From emotion perception to emotion experience:

Emotions evoked by pictures and classical music. International Journal of Psychophysiology. 60, 1 (Apr.

2006), 34–43. doi: 10.1016/j.ijpsycho.2005.04.007

Berndt, D. and Clifford, J. (1994). Using dynamic time warping to find patterns in time series. KDD

workshop (1994), 359–370.

Bianchi, G. and Sorrentino, R. (2007). Electronic filter simulation & design. McGraw-Hill New York.

Buza, K., Koller, J. and Marussy, K. (2015). PROCESS: Projection-based classification of

electroencephalograph signals. Proceedings of ICAISC’15, 14th International Conference on Artificial

Intelligence and Soft Computing, Part II. Lecture Notes in Artificial Intelligence 9120, pp. 91–100,

Springer

Page 22: Emotion Perception and Recognition: An Exploration of Cultural Differences and ... · 2018. 6. 27. · Emotion recognition is one of the most challenging tasks ... examined, demonstrating

Chen L, Ng R (2004). On The Marriage of Lp-norms and Edit Distance. In: Nascimento MA, Özsu MT,

Kossmann D, et al. (eds) Proceedings of the Thirtieth International Conference on Very Large Data Bases,

Toronto, Canada, August 31 - September 3 2004. Morgan Kaufmann, pp 792–803

Chi, Y.M., Wang, Y.T., Wang, Y., Maier, C., Jung, T.P., Cauwenberghs, G., (2012). Dry and Noncontact EEG

Sensors for Mobile Brain–Computer Interfaces, IEEE Trans. Neural Syst. Rehabil. Eng. 20 228–235.

doi:10.1109/TNSRE.2011.2174652

Childers, D.G., Skinner, D.P. and Kemerait, R.C. (1997). The cepstrum: A guide to processing. Proceedings

of the IEEE, vol. 65, no. 10, pp. 1428–1443.

Coan, J.A. and Allen, J.J.B. (2004). Frontal EEG asymmetry as a moderator and mediator of emotion.

Biological Psychology. 67, 1–2 (Oct. 2004), 7–50. doi: 10.1016/j.biopsycho.2004.03.002

Daly, I., Williams, D., Hallowell, J., Hwang, F., Kirke, A., Malik, A. et al., (2015). Music-induced emotions

can be predicted from a combination of brain activity and acoustic features, Brain Cogn. 101, 1–11.

doi:10.1016/j.bandc.2015.08.003

Ding, H., Trajcevski, G., Scheuermann, P., Wang, X. and Keogh, E. (2008). Querying and mining of time

series data: experimental comparison of representations and distance measures. Proceedings of the

VLDB Endowment. 1, 2 (Aug. 2008), 1542–1552. DOI= http://dx.doi.org/10.14778/1454159.1454226

FFmpeg homepage: https://ffmpeg.org/. Accessed: 2015-02-10. Horlings, R. (2008). Emotion recognition

using brain activity. Master’s thesis, Delft University of Technology, Delft, the Netherlands

Huang, N.E., Shen, Z., Long, S.R., Wu, M.C., Shih, H.H., Zheng, Q. et al., (1998) The empirical mode

decomposition and the Hilbert spectrum for nonlinear and non-stationary time series analysis, Proc. R.

Soc. A Math. Phys. Eng. Sci. 454, 903–995. doi:10.1098/rspa.1998.0193

Huang, Y.-J., Wu, C.-Y., Wong, A.M.-K., Lin, B.-S. (2015). Novel Active Comb-Shaped Dry Electrode for

EEG Measurement in Hairy Site, IEEE Trans. Biomed. Eng. 62 256–263. doi:10.1109/TBME.2014.2347318

Iacoviello, D., Petracca, A., Spezialetti, M., Placidi, G. (2015) A real-time classification algorithm for EEG-

based BCI driven by self-induced emotions, Comput. Methods Programs Biomed. 122, 293–303.

doi:10.1016/j.cmpb.2015.08.011

Ivanovic, M., Budimac, Z., Radovanovic, M., Kurbalija, V., Dai, W., Bădică, C. et al., (2015). Emotional

agents - state of the art and applications, Comput. Sci. Inf. Syst. 12, 1121–1148.

doi:10.2298/CSIS141026047I

Jenke, R., Peer, A., Buss, M. (2014). Feature Extraction and Selection for Emotion Recognition from EEG,

IEEE Trans. Affect. Comput. 5, 327–339. doi:10.1109/TAFFC.2014.2339834

Jurcak, V., Tsuzuki, D. and Dan, I. (2007). 10/20, 10/10, and 10/5 systems revisited: Their validity as

relative head-surface-based positioning systems. NeuroImage. 34, 4 (Feb. 2007), 1600–1611. DOI=

http://dx.doi.org/10.1016/j.neuroimage.2006.09.024

Page 23: Emotion Perception and Recognition: An Exploration of Cultural Differences and ... · 2018. 6. 27. · Emotion recognition is one of the most challenging tasks ... examined, demonstrating

Koelstra, S., Muhl, C., Soleymani, M., Lee, J.-S., Yazdani, A., Ebrahimi, T. et al., (2012) DEAP: A Database

for Emotion Analysis Using Physiological Signals, IEEE Trans. Affect. Comput. 3, 18–31. doi:10.1109/T-

AFFC.2011.15

Kurbalija, V., Radovanović, M., Geler, Z. and Ivanović, M. (2010). A Framework for Time-Series Analysis.

Artificial Intelligence: Methodology, Systems, and Applications SE - 5. D. Dicheva and D. Dochev, eds.

Springer Berlin Heidelberg. 42–51. DOI= 10.1007/978-3-642-15431-7_5

Kurbalija V., Radovanović M., Geler Z., Ivanović M. (2011). The influence of global constraints on DTW

and LCS similarity measures for time-series databases. Advances in Intelligent and Soft Computing (Vol

101), Bourgas, Bulgaria, 2011. Springer, pp 67–74.

Kurbalija, V., Radovanović, M., Geler, Z. and Ivanović, M. (2014). The influence of global constraints on

similarity measures for time-series databases. Knowledge-Based Systems. 56, (2014), 49–67. DOI=

http://dx.doi.org/10.1016/j.knosys.2013.10.021

Kurbalija, V., Ivanović, M., Radovanović, M., Geler, Z., Mitrović, D., Dai W. and Zhao W. (2015). Cultural

differences and similarities in emotion recognition. In Proceedings of BCI’15, 7th Balkan Conference in

Informatics, pages 34:1–34:6, Craiova, Romania.

Langner, O., Dotsch, R., Bijlstra, G., Wigboldus, D.H.J., Hawk, S.T., van Knippenberg, A. (2010)

Presentation and validation of the Radboud Faces Database, Cogn. Emot. 24 1377–1388.

doi:10.1080/02699930903485076

Liu, Y., Sourina, O. and Nguyen, M.K. (2010). Real-time EEG-based human emotion recognition and

visualization. Cyberworlds (CW), 2010 International Conference on (2010), 262–269.

doi:10.1109/CW.2010.37

Mikhail, M., El-Ayat, K., El Kaliouby, R., Coan, J. and Allen, J.J.B. (2010). Emotion Detection Using Noisy

EEG Data. Proceedings of the 1st Augmented Human International Conference (New York, NY, USA,

2010), 7:1–7:7. DOI= 10.1145/1785455.1785462

Murugappan, M., Ramachandran, N., Sazali, Y. and others (2010). Classification of human emotion from

EEG using discrete wavelet transform. Journal of Biomedical Science and Engineering. 3, 4 (2010), 390–

396. doi: 10.4236/jbise.2010.34054

Myers, C.S. and Rabiner, L.R. (1981). A comparative study of several dynamic time-warping algorithms

for connected word recognition. The Bell System Technical Journal, vol. 60, no. 7, pp. 1389–1409.

Newberg, L. A. (2009). Error statistics of hidden Markov model and hidden Boltzmann model results.

BMC Bioinformatics. 2009; 10: 212.

Nie, D., Wang, X.-W., Shi, L.-C. and Lu, B.-L. (2011). EEG-based emotion recognition during watching

movies. Neural Engineering (NER), 2011 5th International IEEE/EMBS Conference on (Apr. 2011), 667–

670. doi:10.1109/NER.2011.5910636

Page 24: Emotion Perception and Recognition: An Exploration of Cultural Differences and ... · 2018. 6. 27. · Emotion recognition is one of the most challenging tasks ... examined, demonstrating

Othman, M., Wahab, A., Karim, I., Dzulkifli, M.A., Alshaikli, I.F.T. (2013) EEG Emotion Recognition Based

on the Dimensional Models of Emotions, Procedia - Soc. Behav. Sci. 97, 30–37.

doi:10.1016/j.sbspro.2013.10.201

Oude Bos, D. (2006). EEG-based emotion recognition: The influence of visual and auditory stimuli.

(Online). Available: http://hmi.ewi.utwente.nl/verslagen/capita-selecta/CS-Oude_Bos-Danny.pdf

Petrantonakis, P.C. and Hadjileontiadis, L.J. (2010a). Emotion Recognition From EEG Using Higher Order

Crossings. Information Technology in Biomedicine, IEEE Transactions on. 14, 2 (Mar. 2010), 186–197.

doi:10.1109/TITB.2009.2034649

Petrantonakis, P.C., Hadjileontiadis, L.J. (2010b) Emotion Recognition from Brain Signals Using Hybrid

Adaptive Filtering and Higher Order Crossings Analysis, IEEE Trans. Affect. Comput. 1, 81–97.

doi:10.1109/T-AFFC.2010.7

Picard, R. W. (1997). Affective Computing. The MIT Press, Cambridge, MA, USA,

doi:10.1037/e526112012-054.

Russell, J.A. 1979. Affective space is bipolar. Journal of personality and social psychology. 37, 3 (1979),

345–356. DOI= http://psycnet.apa.org/doi/10.1037/0022-3514.37.3.345

Russell, J.A. (1980) A circumplex model of affect., J. Pers. Soc. Psychol. 39, 1161–1178.

doi:10.1037/h0077714

Salovey, P. and Mayer, J.D. (1990). Emotional intelligence. Imagination, cognition and personality. 9, 3

(1990), 185–211. DOI= 10.2190/DUGG-P24E-52WK-6CDG

Scherer, K.R. (2005) What are emotions? And how can they be measured?, Soc. Sci. Inf. 44, 695–729.

doi:10.1177/0539018405058216

Shahabi, H., Moghimi, S. (2016). Toward automatic detection of brain responses to emotional music

through analysis of EEG effective connectivity, Comput. Human Behav. 58, 231–239.

doi:10.1016/j.chb.2016.01.005

Shenoi, B.A. (2005). Introduction to digital signal processing and filter design. John Wiley & Sons.

Society, A.C.N. 2006. (2006). Guideline 5: Guidelines for standard electrode position nomenclature.

American Journal of Electroneurodiagnostic Technology. 46, 3, 222–225. DOI=

10.1080/1086508X.2006.11079580

Tao, J. and Tieniu, T. (2005). Affective computing: A review. Affective Computing and Intelligent

Interaction. Lecture Notes in Computer Science 3784, Springer, pp. 981–995.

Tomašev, N., Buza, K., Marussy, K. and Kis, P. B. (2015). Hubness-aware classification, instance selection

and feature construction: Survey and extensions to time-series. U. Stańczyk and C. L. Jain (eds.), Feature

Page 25: Emotion Perception and Recognition: An Exploration of Cultural Differences and ... · 2018. 6. 27. · Emotion recognition is one of the most challenging tasks ... examined, demonstrating

Selection for Data and Pattern Recognition. Studies in Computational Intelligence 584, pp. 231–262,

Springer

Wang, X.-W., Nie, D., Lu, B.-L. (2014). Emotional state classification from EEG data using machine

learning approach, Neurocomputing. 129, 94–106, doi: 10.1016/j.neucom.2013.06.046

Yoon, H.J., Chung, S.Y. (2013) EEG-based emotion estimation using Bayesian weighted-log-posterior

function and perceptron convergence algorithm, Comput. Biol. Med. 43, 2230–2237.

doi:10.1016/j.compbiomed.2013.10.017

Zheng, W.-L., Lu, B.-L. (2015). Investigating Critical Frequency Bands and Channels for EEG-Based

Emotion Recognition with Deep Neural Networks, IEEE Trans. Auton. Ment. Dev. 7 162–175.

doi:10.1109/TAMD.2015.2431497