1
Sentence classification based on phase patterns in EEG neural oscillation during imagined speech Hiroki WATANABE, Hiroki TANAKA, Sakriani SAKTI, Satoshi NAKAMURA 調音動作想像時の脳波位相パターンを用いた文識別 Email: [email protected] Research questions Analysis pipeline speech rhythm (4-8Hz) neural oscillation in STG (4-8Hz) phase synchronization Speech comprehension phase synchronization b/w produced speech & neural oscillation in the motor area? Hypothesis: imagined speech speech rhythms replicable oscillation patterns spoken sentence classification (Watanabe et al., 2017) RQ. 1 Do EEG phase patterns during imagined speech synchronize with produced speech rhythms? RQ.2 How accurately do EEG phase patterns classify imagined sentences? EEG experiment Participant 1 male, 1 female (L1: Japanese) Speech material 1. あなたが昨日夢中で読んでいた本はおもしろかった。 (The book that you were absorbed in yesterday is interesting.) 2. ついさっき女の子が私に言ったことは本当の話。 (What the girl said to me just now is true.) 3. 向こうの壁に飾っているのは彼のお兄さんが描いた絵。 (The picture on the other wall was drawn by his older brother.) Tasks feature extraction model training evaluation phase-locking value (PLV) PLV(ch) = 1 =1 exp(1 ∗ − ) preprocess perception data: 35 trials (subj 01), 38 trials (subj 02) imagined data: 39 trials (subj 01), 42 trials (subj 02) perception: fronto-central imagined: centro-parietal Averaged across target channels FFT decomposition Phase extraction: [0, 2.8] sec, [2, 12] Hz Logistic regression SVM Leave-one-out cross-validation Introduction Indication of a task execution Ready? LISTEN + + + SPEAK + + + IMAGINE + + + 2 sec 2.5 sec 5.5 sec 2.5 sec listening model speech is played speaking imitating model speech orally imaging imaging articulatory movement as the same rate to the speaking task Research questions Conclusion & Future directions RQ. 1 A tendency of phase synchronization b/w produced speech & EEG oscillations during imagined sentences RQ. 2 Best accuracies: [52.4 59.0 %] by SVM. Phase patterns are reliable features for imagined sentence recognition Future directions : (1) Increasing number of participants + (2) Source localization of this phase synchronization phenomenon Investigating… (1) a phase synchronization during imaging articulatory movements of sentences (2) accuracies of EEG-based imagined sentence classification using the phase synchronization as features Research purposes If there is a synchronization, phase information is useful features for classification bandpass-filter (1-14Hz) speech data EEG speech phase EEG phase perception: model imagined: produced perception: listening task imagined: imagined task Synchronization analysis Classification analysis: 3 sentence classification task Selected references H. Luo and D. Poeppel. (2007) Phase patterns of neuronal responses reliably discriminate speech in human auditory cortex, Neuron, vol.5, pp.1001–1010. M. D ’Zmura, S. Deng, T.L.S. Thorpe, and R. Srinivasan. (2009) Toward EEG sensing of imagined speech, In Human-Computer Interaction. New Trends, Springer, Berlin, Heidelberg, pp.40–48. C.S. DaSalla, H. Kambara, M. Sato, and Y. Koike. (2009) Single-trial classification of vowel speech imagery using common spatial patterns, Neural Networks, vol.22, no.9, pp.1334–1339. H. Watanabe, H. Tanaka, S. Sakti, and S. Nakamura. (2017) Subject-independent classification of Japanese spoken sentences by multiple frequency bands phase pattern of EEG response during speech perception, In Proceedings of Interspeech, Stockholm, Sweden, pp.2431–2435. speech imagination in the head Feature extraction How are you? Measuring EEG Model training Estimation of speech How are you? A big picture of our research Graduate School of Information Science, Nara Institute of Science and Technology, Nara, Japan Classification analysis subj 01 percep: 40.0 % (rf) imagine: 59.0 % (log reg/ SVM) subj 02 percep: 57.9 % (logreg/template) imagine: 52.4 % (SVM) logreg rf SVM template Result Averaged PLV across two participants perception imagined Synchronization analysis Perception: stronger PLV in fronto/ central region Imagined: stronger PLV in centro/ parietal region logistic regression: logreg random forest: rf template matching: template Accuracy (%) Accuracy (%) imagined perception imagined perception logreg rf SVM template Random forest Template matching Previous research & research focus EEG-based sentence classification (Watanabe et al, 2017) vowel classification (DaSalla et al., 2009) syllable classification (D’zmura et al., 2009) word classification (Martin et al., 2015) classification of imagined speech sentence classification no research Applying an EEG-based sentence classification method to imagined speech recognition classification of heard speech Utilizing a phase synchronization to sentence rhythms

Sentence classification based on phase patterns in … › papers › conference › 2018 › 201807...Sentence classification based on phase patterns in EEG neural oscillation during

  • Upload
    others

  • View
    11

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Sentence classification based on phase patterns in … › papers › conference › 2018 › 201807...Sentence classification based on phase patterns in EEG neural oscillation during

Sentence classification based on phase patterns in EEG neural oscillation during imagined speech

Hiroki WATANABE, Hiroki TANAKA, Sakriani SAKTI, Satoshi NAKAMURA

調音動作想像時の脳波位相パターンを用いた文識別

Email: [email protected]

Research questions

Analysis pipeline

ー speech rhythm (4-8Hz) ー neural oscillation in STG (4-8Hz)

phase synchronization

Speech comprehension

phase synchronization b/w produced speech & neural oscillation in the motor area?

Hypothesis: imagined speech

speech rhythms ➡ replicable oscillation patterns

spoken sentence classification (Watanabe et al., 2017)

RQ. 1 Do EEG phase patterns during imagined speech synchronize with produced speech rhythms?

RQ.2 How accurately do EEG phase patterns classify imagined sentences?

EEG experimentParticipant 1 male, 1 female (L1: Japanese)Speech material

1. あなたが昨日夢中で読んでいた本はおもしろかった。(The book that you were absorbed in yesterday is interesting.)

2. ついさっき女の子が私に言ったことは本当の話。(What the girl said to me just now is true.)

3. 向こうの壁に飾っているのは彼のお兄さんが描いた絵。(The picture on the other wall was drawn by his older brother.)

Tasks

feature extraction

modeltraining

evaluation

phase-locking value (PLV)

PLV(ch) =

1

𝑇

𝑡=1

𝑇

exp(1𝑖 ∗ 𝐸𝐸𝐺𝜑 − 𝑆𝑃𝐸𝐸𝐶𝐻𝜑 )

preprocess• perception data: 35 trials (subj 01), 38 trials (subj 02)• imagined data: 39 trials (subj 01), 42 trials (subj 02)

perception: fronto-centralimagined: centro-parietal

• Averaged across target channels• FFT decomposition• Phase extraction: [0, 2.8] sec, [2, 12] Hz

• Logistic regression• SVM

• Leave-one-out cross-validation

Introduction

Indication of a task execution

Ready? LISTEN + + +

SPEAK + + +

IMAGINE + + +

2 sec 2.5 sec 5.5 sec 2.5 sec

listeningmodel speech is played

speakingimitating model speech orally

imagingimaging articulatory movementas the same rate to the speakingtask

Research questions

Conclusion & Future directionsRQ. 1 A tendency of phase synchronization b/w produced speech & EEG

oscillations during imagined sentencesRQ. 2 Best accuracies: [52.4 59.0 %] by SVM.

Phase patterns are reliable features for imagined sentence recognition

Future directions: (1) Increasing number of participants +

(2) Source localization of this phase synchronization phenomenon

Investigating…(1) a phase synchronization during imaging articulatory movements of sentences(2) accuracies of EEG-based imagined sentence classification using the phase synchronization as features

Research purposes

If there is a synchronization, phase information is useful features for classification

bandpass-filter (1-14Hz)

speech data EEG

speech phase EEG phase

perception: modelimagined: produced

perception: listening task imagined: imagined task

Synchronization analysis

Classification analysis: 3 sentence classification task

Selected referencesH. Luo and D. Poeppel. (2007) Phase patterns of neuronal responses reliably discriminate speech in human auditory cortex, Neuron, vol.5, pp.1001–1010.M. D ’Zmura, S. Deng, T.L.S. Thorpe, and R. Srinivasan. (2009) Toward EEG sensing of imagined speech, In Human-Computer Interaction. New Trends, Springer, Berlin,

Heidelberg, pp.40–48.C.S. DaSalla, H. Kambara, M. Sato, and Y. Koike. (2009) Single-trial classification of vowel speech imagery using common spatial patterns, Neural Networks, vol.22, no.9,

pp.1334–1339.H. Watanabe, H. Tanaka, S. Sakti, and S. Nakamura. (2017) Subject-independent classification of Japanese spoken sentences by multiple frequency bands phase pattern of

EEG response during speech perception, In Proceedings of Interspeech, Stockholm, Sweden, pp.2431–2435.

speech imaginationin the head

Feature extraction

How are you?

Measuring EEG

Model training

Estimation of speech

How are you?

A big picture of our research

Graduate School of Information Science, Nara Institute of Science and Technology, Nara, Japan

Classification analysissubj 01

percep: 40.0 % (rf)imagine: 59.0 % (log reg/ SVM)

subj 02percep: 57.9 % (logreg/template) imagine: 52.4 % (SVM)

logreg rf SVM template

Result

Averaged PLV across two participants

perception imaginedSynchronization analysis

Perception: stronger PLV infronto/ central region

Imagined: stronger PLV incentro/ parietal region

logistic regression: logregrandom forest: rftemplate matching: template

Acc

ura

cy (

%)

Acc

ura

cy (

%)

■ imagined■ perception

■ imagined■ perception

logreg rf SVM template

• Random forest• Template matching

Previous research & research focus

• EEG-based sentence classification(Watanabe et al, 2017)

• vowel classification (DaSalla et al., 2009)

• syllable classification (D’zmura et al., 2009)

• word classification (Martin et al., 2015)

classification of imagined speech

• sentence classification ➡ no research

Applying an EEG-based sentence classification method to imagined

speech recognition

classification of heard speech

• Utilizing a phase synchronization to sentence rhythms