The software could support a number of developments, including helping the hard of hearing to navigate the world around them. The machine’s mistakes were small, including things like missing an “s” at the end of a word, or single letter misspellings. The human lip-reader correctly read 12 per cent of words, while the WAS software recognised 50 per cent of the words in the dataset, without error. They found that the software system was more accurate compared to the professional. Researchers from the Oxford University in the UK compared the ability of the machine and a human expert to work out what was being said in the silent video by focusing solely on each speaker’s lip movements. The videos contained more than 118,000 sentences in total, and a vocabulary of 17,500 words. The Watch, Attend and Spell (WAS) software system uses computer vision and machine learning methods to learn how to lip-read from a dataset made up of more than 5,000 hours of TV footage, gathered from six different programmes. Oxford scientists have developed a new artificial intelligence programme that can lip-read more accurately than people, an advance that will help those who suffer from hearing loss.
0 Comments
Leave a Reply. |