In the domain of Human-Computer Interaction (HCI), the main focus of the computer is to interpret the external stimuli provided by users. Moreover in the multi-person scenar- ios, it is important to localize and track the speaker. To solve this issue, we introduce here a framework by which multi-modal sensory data can be eciently and meaningfully com- bined in the application of speaker tracking. This framework fuses together four dierent observation types taken from multi-modal sensors. The advantages of this fusion are that weak sensory data from either modality can be reinforced, and the presence of noise can be reduced. We propose a method of combining these modalities by employing a particle lter. This method oers satised real-time performance. We demonstrate results of a speaker localization in two- and three-person scenarios.
목차
Abstract 1: Introduction 2: The Proposed Approach 2.1: Video Modality 2.2: Audio Modality 2.3: Particle Filter Implementation 3: Experimental Results 4: Conclusions and Future Work Acknowledgments References
보안공학연구지원센터(IJFGCN) [Science & Engineering Research Support Center, Republic of Korea(IJFGCN)]
설립연도
2006
분야
공학>컴퓨터학
소개
1. 보안공학에 대한 각종 조사 및 연구
2. 보안공학에 대한 응용기술 연구 및 발표
3. 보안공학에 관한 각종 학술 발표회 및 전시회 개최
4. 보안공학 기술의 상호 협조 및 정보교환
5. 보안공학에 관한 표준화 사업 및 규격의 제정
6. 보안공학에 관한 산학연 협동의 증진
7. 국제적 학술 교류 및 기술 협력
8. 보안공학에 관한 논문지 발간
9. 기타 본 회 목적 달성에 필요한 사업
간행물
간행물명
International Journal of Future Generation Communication and Networking
간기
격월간
pISSN
2233-7857
수록기간
2008~2016
십진분류
KDC 505DDC 605
이 권호 내 다른 논문 / International Journal of Future Generation Communication and Networking Vol.5 No.4