2000 character limit reached
Understanding the visual speech signal (1710.01351v1)
Published 3 Oct 2017 in cs.CV and eess.AS
Abstract: For machines to lipread, or understand speech from lip movement, they decode lip-motions (known as visemes) into the spoken sounds. We investigate the visual speech channel to further our understanding of visemes. This has applications beyond machine lipreading; speech therapists, animators, and psychologists can benefit from this work. We explain the influence of speaker individuality, and demonstrate how one can use visemes to boost lipreading.
- Helen L Bear (9 papers)