Decoding visemes: Improving machine lip-reading
Bear, Y. and Harvey, Richard 2016. Decoding visemes: Improving machine lip-reading. in: 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) IEEE.
|Authors||Bear, Y. and Harvey, Richard|
To undertake machine lip-reading, we try to recognise speech from a visual signal. Current work often uses viseme classification supported by language models with varying degrees of success. A few recent works suggest phoneme classification, in the right circumstances, can outperform viseme classification. In this work we present a novel two-pass method of training phoneme classifiers which uses previously trained visemes in the first pass. With our new training algorithm, we show classification performance which significantly improves on previous lip-reading results.
|Keywords||visemes; weak learning; visual speech; lip-reading; recognition; classification|
|Book title||2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)|
|19 May 2016|
|Publication process dates|
|Deposited||24 Feb 2017|
|Event||The 41st IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)|
|Digital Object Identifier (DOI)||doi:10.1109/ICASSP.2016.7472029|
|Web address (URL)||http://ieeexplore.ieee.org/document/7472029/|
© 2016 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.
|Accepted author manuscript|
1views this month
7downloads this month