ahybridsystemsapproachto …hiroaki/publication/... · s s r t b ; ^ \ q u k u | \ w q v | g ø r w...

6
ϋΠϒϦουγεςϜʹΑΔ ϚϧνϝσΟΞɾλΠϛϯάߏͷϞσϧԽͱश ژେใڀݚՊ ˟ 606–8501 ژژࠨࢢ٢ຊொ E-mail: [email protected] Β· ΧϝϥϚΠΫͱηϯαΒಘΒΕΔܥσʔλΒɼքʹΔਓͷಈɼදมԽɼ ͳͲͷಈతΛϞσϧԽɾղੳΔख๏ͱɼຊڀݚͰʮϋΠϒϦουγεςϜʯɼಛʹɼෳͷܥΛ ҐͷܥʢΦʔτϚτϯʣʹΑΓସΔʹܥΔɽͷͱɼରͷμΠφϛΫεͷ ୯७ͳϓϦϛςΟϒͰදΕɼΕΒͷΓସΘΓͱදݱͰΔ߹ͳͲʹɼϞσϧϕʔεͰͷΫϥελϦϯ άΛߦͱͰɼΕΒϓϦϛςΟϒΛಈతʹઅԽͳΒಈతϞσϧͱਪఆΛߦͱͰΔɽΒʹɼ ΜಉఆΕγεςϜΒͼशσʔλʹܥ৴߸ΛੜΔͱՄͰΔɽຊߘͰΕΒͷ ಛΛԠ༻දղੳෳϝσΟΞ৴߸ͷλΠϛϯάߏϞσϧΛհΔɽ Ωʔϫʔυ ϋΠϒϦουγεςϜɼઢܗಈతγεςϜɼλΠϛϯάɼදɼௌ౷߹ A Hybrid Systems Approach to Modeling and Learning Multimedia Timing Structures Hiroaki KAWASHIMA Graduate School of Informatics, Kyoto University Yoshida-honmachi, Sakyo, Kyoto, 606–8501 Japan E-mail: [email protected] Abstract Capturing dynamic events of human body motion, facial action, and speech, via sensors, e.g., cameras and microphones, we can obtain a variety of temporal data. To model and analyze the dynamics of captured targets from the data, we introduce a method to utilize ”hybrid systems.” We here use a particular model of hybrid systems that switches several dynamical systems by the upper-level discrete-event system (automaton). If target dynamics can be represented by a set of simple primitives, a model-based clustering algorithm can automatically segment and estimate each of the primitive dynamic models. Besides, an identified model can generate signals similar to given training data. In this report, we show some applications that utilize these features, such as facial-expression analysis and multimedia timing modeling. Key words Hybrid system, linear dynamical system, timing, facial expression, audio-visual integration 1. Ίʹ ਓͷදɼਓಉͷίϛϡχέʔγϣϯΛɼΧϝϥ ϚΠΫͳͲͷηϯαͰܭଌΔͱɼʹʮλΠϛϯάʯ ʹΔʑͳใଘΔɽΑղੳΕΔͷͱɼ ͷλʔϯςΠΩϯάͷύλʔϯͷ߹ʢସજʣ Δɽ·ಈͷϦζϜςϯϙͳͲͷใɼ ͷҙຯΛཧղɾղऍΔͰॏཁͰΔɼදͷมԽͻͱ ͱɼݩޱͱݩͷมԽ։ͷͳͲɼͳ ߏӅΕΔɽͷΑͳʮλΠϛϯάʯͷใΛදݱɼ ΔΊͷϞσϧͱɼຊߘͰϋΠϒϦουγεςϜ ͱݺΕΔཧϞσϧΛհΔɽ ηϯαͷܭଌσʔλʹ·ΕΔಈతͳΛݕग़ɼΔ ͰɼܥύλʔϯΛѻʑͳख๏ఏҊΕΔɽ ͱಈతܭը๏Λ༻ϚονϯάΛߦ๏ɼhidden Markov model (HMM) ΧϧϚϯϑΟϧλͳͲͷܥύ λʔϯͷϞσϧΛಋೖΔ๏ͳͲΔ [1]ɽͷɼ ܥύλʔϯͷΫϥεมಈΛදݱΔΊʹɼܥύ λʔϯͷϞσϧΛ༻Δͱ༗ͰΓɼʹ 80 ޙΒɼಈతͷʑͳϞσϧԽख๏ఏҊΕΔɽ Λ߸هతʹѻϞσϧͱɼ༗ݶঢ়ଶΦʔτϚτϯ ϖτϦωοτͷΑͳ ܥ(discrete-event system) − 63 − 一般社団法人 電子情報通信学会 信学技報 THE INSTITUTE OF ELECTRONICS, IEICE Technical Report INFORMATION AND COMMUNICATION ENGINEERS Copyright ©2014 by IEICE This article is a technical report without peer review, and its polished and /or extended version may be published elsewhere. SIS2014-78 (2014-12)

Upload: others

Post on 17-Jul-2020

0 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: AHybridSystemsApproachto …hiroaki/publication/... · s s r t b ; ^ \ q U K U | \ w q V | G ø r w O t [ b T x è × p ) Q \ q t s } h i ` | î M t x ô ø q G ø q r w O t 0 ^ d

THE INSTITUTE OF ELECTRONICS,INFORMATION AND COMMUNICATION ENGINEERS

TECHNICAL REPORT OF IEICE.

† 606–8501

E-mail: †[email protected]

A Hybrid Systems Approach to

Modeling and Learning Multimedia Timing Structures

Hiroaki KAWASHIMA†

† Graduate School of Informatics, Kyoto University Yoshida-honmachi, Sakyo, Kyoto, 606–8501 Japan

E-mail: †[email protected]

Abstract Capturing dynamic events of human body motion, facial action, and speech, via sensors, e.g., cameras

and microphones, we can obtain a variety of temporal data. To model and analyze the dynamics of captured targets

from the data, we introduce a method to utilize ”hybrid systems.” We here use a particular model of hybrid systems

that switches several dynamical systems by the upper-level discrete-event system (automaton). If target dynamics

can be represented by a set of simple primitives, a model-based clustering algorithm can automatically segment

and estimate each of the primitive dynamic models. Besides, an identified model can generate signals similar to

given training data. In this report, we show some applications that utilize these features, such as facial-expression

analysis and multimedia timing modeling.

Key words Hybrid system, linear dynamical system, timing, facial expression, audio-visual integration

1.

hidden

Markov model (HMM)

[1]

80

(discrete-event system)

— 1 —− 62 − − 63 −

一般社団法人 電子情報通信学会 信学技報THEINSTITUTEOFELECTRONICS, IEICETechnicalReportINFORMATIONANDCOMMUNICATIONENGINEERS

Copyright ©2014 by IEICEThis article is a technical report without peer review, and its polished and /or extended version may be published elsewhere.

SIS2014-78(2014-12)

Page 2: AHybridSystemsApproachto …hiroaki/publication/... · s s r t b ; ^ \ q U K U | \ w q V | G ø r w O t [ b T x è × p ) Q \ q t s } h i ` | î M t x ô ø q G ø q r w O t 0 ^ d

(dynamical system)

(hybrid dynamical system, HDS)

(switched dynamical

system, switching dynamical system)

[2] [3] [4]

HDS

HDS

3.

HDS

HDS

Interval-Based Hybrid Dynamical System

),|,( 1 pikjk qIqIP

Linear dynamical systems

)()(1

)( it

it

it wgxFx

ttt vHxy

State transition

Observation

,jqduration

Intervals

time <q1, 3> <q1, 2> <q2, 4>

q1 q2

P( <q1, 1> | <q2, 2> )

Finite state automaton P( <q2, 2> | <q1, 1> )

time

Internal state sequence Observation H

Signal generation

Internal state space

Dynamical System D2

Dynamical System D1

Interval sequence

Interval-based transition Discrete state

1 Interval hybrid dynamical system

2.

2. 1

HDS 1

(linear dynamical system, LDS)

1

D = {D1, ..., DN}Q = {q1, ..., qN}

qi Di

2. 2

t xt ∈ Rn yt ∈ Rny

Di

xt = A(i)xt−1 + b(i) + ω(i)t (1)

yt = Cxt + υt, (2)

ω(i) υ

0 1 Di

2

A(i) b(i)

C .

Di A(i), b(i)

ω(i)t

2. 3

— 2 —

qi τ

⟨qi, τ⟩

P (Ik = ⟨qj , τ⟩|Ik−1 = ⟨qi, τp⟩), (3)

1

I1, ..., IK

P (qj |qi) P (τ |τp, qi, qj)

segment model [2]

HDS

(interval HDS, IHDS)

2. 4

x0

y0, y1, ... IHDS

xt

IHDS

3.

HDS

IHDS

expectation-maximization (EM)

EM

1 Ik 2

k

Training data set

各LDSのパラメタをrefine 確率オートマトンのパラメタ推定

EMによるrefinementプロセス

LDSの数を決定 各LDSの大まかなパラメタを推定

初期化

LDS集合の階層的クラスタリング

Typical data set

fix

fix

2 Two-stage learning method

EM

2

[5]

1

Kullback-Leibler divergence

2 EM

1

EM 1

EM

EM

Viterbi

E IHDS

2. 4

M IHDS

4. 1

Ekman Facial

Action Coding System (FACS) [6]

Action Unit

— 3 —− 64 − − 65 −

Page 3: AHybridSystemsApproachto …hiroaki/publication/... · s s r t b ; ^ \ q U K U | \ w q V | G ø r w O t [ b T x è × p ) Q \ q t s } h i ` | î M t x ô ø q G ø q r w O t 0 ^ d

qi τ

⟨qi, τ⟩

P (Ik = ⟨qj , τ⟩|Ik−1 = ⟨qi, τp⟩), (3)

1

I1, ..., IK

P (qj |qi) P (τ |τp, qi, qj)

segment model [2]

HDS

(interval HDS, IHDS)

2. 4

x0

y0, y1, ... IHDS

xt

IHDS

3.

HDS

IHDS

expectation-maximization (EM)

EM

1 Ik 2

k

Training data set

各LDSのパラメタをrefine 確率オートマトンのパラメタ推定

EMによるrefinementプロセス

LDSの数を決定 各LDSの大まかなパラメタを推定

初期化

LDS集合の階層的クラスタリング

Typical data set

fix

fix

2 Two-stage learning method

EM

2

[5]

1

Kullback-Leibler divergence

2 EM

1

EM 1

EM

EM

Viterbi

E IHDS

2. 4

M IHDS

4. 1

Ekman Facial

Action Coding System (FACS) [6]

Action Unit

— 3 —− 64 − − 65 −

Page 4: AHybridSystemsApproachto …hiroaki/publication/... · s s r t b ; ^ \ q U K U | \ w q V | G ø r w O t [ b T x è × p ) Q \ q t s } h i ` | î M t x ô ø q G ø q r w O t 0 ^ d

100 500

280 300 320 410 430 460

l. eye

frame#

frame#

r. eye

nose

mouth

Original images

in the sequence

Obtained

facial score

3 (Facial Score)

neutral smiling

0 0.5time[s]

begin

smilingneutral smiling

time[s]

begin

smiling

0 0.5

left eye

nose

mouth

subject A

subject B

0

20

40

60

80

100

-40 -20 0 20 40 60Mb(e

nose -

bm

outh

) [

fram

e]

Me(eleye - emouth) [frame]

-40

-30

-20

-10

0

10

0 10 20 30 40 50Mb(e

no

se -

em

ou

th)

[fr

am

e]

Mb(eleye - bnose) [frame]

intentional smilespontaneous smile

intentional smilespontaneous smile

4 [9]

[7]

4. 1

active appearance model

(AAM) [8]

3. IHDS

3

4. 2

2

Mb Me

4( )

2

2

a2 a3

b1 b2

a1Signal A

Signal B b3

a2 a3

b1 b2

a1

b3

Difference of beginning points

(beg(a3) - beg(b2))

Difference of ending points

(end(a3) - end(b2))

a1

Temporal difference distribution of pair ( )a3 b2

5 [11]

leave-one-out

support vector machine

6

80-100% 79-96%

5. 2

HMM

[10]

/pa/

/a/

[11]

5. 1

IHDS

2

2 2

5

[11]

[12]

5. 2

2 Sa, Sb

— 4 —

Training pair of interval sequences

Generated image sequence

Original image sequence

Audio interval sequence

Visual interval sequence

Generated visual interval sequence from audio interval sequenceTime [frame]

Mode #1

#8

Frame #140 #250

-0.2

-0.1

0

0.1

0.2

Reference (input) audio signal

6 [11]

Sa Sb

1 Sa I(a) = [I(a)1 , ..., I

(a)Ka]

2 Sa I(a)

Sb I(b) = [I(b)1 , ..., I

(b)Kb]

3 I(b) Sb

(1) (3)

IHDS

2. 4

(2)

Sa I(a)

I(b)

5. 3

/a//i//u//e//o/

9

6 1,2

I(a) ,

I(b) 6 3

I(b)

IHDS

.

140 250 5

6 5 6

6

Speech

Spectrum analysis

Visual feature extraction

Face detection Lip tracking

Step B: Consistency evaluation

Step A: Candidate generation

X=S+N

Sc1, Sc2, …

V

Estimated clean speech

S

Video

7

[12]

Audio only

Proposed

Separated Noise

Estimated Speech

Estimated Speech

8 (2 ) 3

1

5. 4

1

7

5. 2

(2)

parallel list Viterbi [13]

1

(3)

IHDS

Sc1, Sc2, ... X

S [14]

8 2, 3

— 5 —− 66 − − 67 −

Page 5: AHybridSystemsApproachto …hiroaki/publication/... · s s r t b ; ^ \ q U K U | \ w q V | G ø r w O t [ b T x è × p ) Q \ q t s } h i ` | î M t x ô ø q G ø q r w O t 0 ^ d

Training pair of interval sequences

Generated image sequence

Original image sequence

Audio interval sequence

Visual interval sequence

Generated visual interval sequence from audio interval sequenceTime [frame]

Mode #1

#8

Frame #140 #250

-0.2

-0.1

0

0.1

0.2

Reference (input) audio signal

6 [11]

Sa Sb

1 Sa I(a) = [I(a)1 , ..., I

(a)Ka]

2 Sa I(a)

Sb I(b) = [I(b)1 , ..., I

(b)Kb]

3 I(b) Sb

(1) (3)

IHDS

2. 4

(2)

Sa I(a)

I(b)

5. 3

/a//i//u//e//o/

9

6 1,2

I(a) ,

I(b) 6 3

I(b)

IHDS

.

140 250 5

6 5 6

6

Speech

Spectrum analysis

Visual feature extraction

Face detection Lip tracking

Step B: Consistency evaluation

Step A: Candidate generation

X=S+N

Sc1, Sc2, …

V

Estimated clean speech

S

Video

7

[12]

Audio only

Proposed

Separated Noise

Estimated Speech

Estimated Speech

8 (2 ) 3

1

5. 4

1

7

5. 2

(2)

parallel list Viterbi [13]

1

(3)

IHDS

Sc1, Sc2, ... X

S [14]

8 2, 3

— 5 —− 66 − − 67 −

Page 6: AHybridSystemsApproachto …hiroaki/publication/... · s s r t b ; ^ \ q U K U | \ w q V | G ø r w O t [ b T x è × p ) Q \ q t s } h i ` | î M t x ô ø q G ø q r w O t 0 ^ d

6.

SCAT

[1] , .

. (2006-CVIM-154), pp.

197–209, 2006.

[2] M. Ostendorf, V. Digalakis, and O. A. Kimball. From

HMMs to segment models: A unified view of stochastic

modeling for speech recognition. IEEE Trans. Speech and

Audio Process, Vol. 4, No. 5, pp. 360–378, 1996.

[3] C. Bregler. Learning and recognizing human dynamics in

video sequences. Proc. Int. Conference on Computer Vision

and Pattern Recognition, pp. 568–574, 1997.

[4] Y. Li, T. Wang, and H.-Y. Shum. Motion texture: A two-

level statistical model for character motion synthesis. Proc.

SIGGRAPH, pp. 465–472, 2002.

[5] H. Kawashima and T. Matsuyama. Multiphase learning for

an interval-based hybrid dynamical system. IEICE Trans.

Fundamentals, Vol. E88-A, No. 11, pp. 3022–3035, 2005.

[6] P. Ekman and W. V. Friesen. Unmasking the Face. Prentice

Hall, 1975.

[7] , , , . :

.

, Vol. 9, No. 2, pp. 201–211, 2007.

[8] T. F. Cootes, G. J. Edwards, and C. J. Taylor. Active ap-

pearance model. Proc. European Conference on Computer

Vision, pp. 484–498, 1998.

[9] , .

. / / , Vol. 54,

No. 1, pp. 28–33, 2010.

[10] M. Brand, N. Oliver, and A. Pentland. Coupled hidden

Markov models for complex action recognition. Proc. IEEE

Conference on Computer Vision and Pattern Recognition,

pp. 994–999, 1997.

[11] , .

.

, Vol. 48, No. 12, pp. 3680–3691, 2007.

[12] Hiroaki Kawashima, Yu Horii, and Takashi Matsuyama.

Speech estimation in non-stationary noise environments us-

ing timing structure between mouth movements and sound

signals. Interspeech, pp. 442–445, 2010.

[13] N. Seshadri and C.-E. W. Sundberg. List Viterbi decoding

algorithms with applications. IEEE Trans. on Communi-

cations, Vol. 42, No. 234, pp. 313–323, 1994.

[14] M. Fujimoto and S. Nakamura. A non-stationary noise sup-

pression method based on particle filtering and polyak aver-

aging. IEICE Trans. on Information and Systems, Vol. 89,

No. 3, pp. 922–930, 2006.

— 6 —

社団法人 電子情報通信学会THE INSTITUTE OF ELECTRONICS,INFORMATION AND COMMUNICATION ENGINEERS

信学技報TECHNICAL REPORT OF IEICE.

多重化不可視映像技術(第 2報)

―FPGAを用いたハードウェア化―

田口 裕起† 鈴木 久貴† 白井 暁彦†

† 神奈川工科大学 〒 243–0292 神奈川県厚木市下荻野 1030E-mail: †{taguchi,hisataka}@shirai.la, ††[email protected]

あらまし 本論文は多重化不可視映像を実現する FPGA画像合成ハードウェアについて報告する.この技術は民生用

パッシブ 3Dディスプレイの新しい付加価値創出技術である,多重化不可視映像生成アルゴリズム「ExPixel」に基

いている.このアルゴリズムは,自然画像とは別に,裸眼には視認できない映像を生成することができるが,同時に

その映像は円偏光フィルターによって選択的に同時に視聴可能である.この技術は,いかなるハードウェア改造や電

子的特殊デバイスを使用せず,物理現象とソフトウェア,高解像度パッシブ 3Dディスプレイによって実装可能であ

る.本論文では ATYLSプラットフォームにおける,2つの HDMIからフル HD解像度で 1つの HDMI出力をする

ExPixel FPGA実装を報告する.本ハードウェアは幅広い機器でのハードウェア接続試験を行い,従来の PCベース

のアプリケーションから,家電,放送,ビデオゲーム機器といった組み込み機器への多重化不可視映像の応用可能性

を拡張した.

キーワード 画像処理,多重化,3D,実時間処理,FPGA,回路設計

Multiplex hidden imagery technology (II)

Implement on FPGA based hardware

Hiroki TAGUCHI†, Hisataka SUZUKI†, and Akihiko SHIRAI†

† Kanagawa Institute of Technology 1030 Shimogino, atsugi–shi, Kanagawa, 243–0292 JapanE-mail: †{taguchi,hisataka}@shirai.la, ††[email protected]

Abstract This article contributes to develop multiplex-hidden image generation hardware implementation on

FPGA. The technology is based on a new use of consumer passive 3D displays. The proposed algorithm, which

named as ExPixel, can be realized hidden image and prior natural image for naked eyes, but it can be shown by a

circular polarization filter selectively and simultaneously. The technique can be implemented in software and phys-

ical phenomenon thanks to high resolution of current passive 3D high definition displays, instead of any modified

hardwares and/or special electronic devices. In this article, we reported a prototype of FPGA implementation of

ExPixel on ATYLS platform with two HDMI input and one HDMI output in full HD resolution. It could tested

on various hardware connectivity and it can enlarge new applications of multiplex hidden imagery, not only PC

applications but also embedded platform like home electronics, broadcasting materials and video game consoles.

Key words Image Processing, Multiplex Image, 3D, Real-time Operation, FPGA, Circuit Design

1. は じ め に

近年,3Dディスプレイの一般化とともに映像表示機器の高

解像度化が進んでおり,人間の知覚限界に近づいてきている.

そのため,表示機器には高解像度化以外の付加価値が求められ

ている.付加価値には様々なアプローチがあり,色域を広げる

ためサブピクセルの色数を 4色としたディスプレイや,液晶の

倍速駆動に対応したディスプレイなどがある.

立体表示対応のディスプレイが 2010年頃から流行している,

しかし,過去の歴史からも立体映像の流行は長期的には続かず

数年で鎮静化する傾向にあるとされている [1].そこで,さらな

る付加価値を与える技術として,2つの異なる映像を同時に表

示する多重化不可視映像技術が提案されている [2].

この技術は,スタック式 3D プロジェクタやパッシブ式 3D

— 1 —− 68 − − 69 −

一般社団法人 電子情報通信学会 信学技報THEINSTITUTEOFELECTRONICS, IEICETechnicalReportINFORMATIONANDCOMMUNICATIONENGINEERS