"audio-visual speech corpus; speech recognition; lip tracking"@en . . . . "6"^^ . "Zelinka, Jan" . . . . . "Czech Audio-Visual Speech Corpus for Recognition with Impaired Conditions" . "Czech Audio-Visual Speech Corpus for Recognition with Impaired Conditions"@en . . . "6"^^ . "M\u00FCller, Lud\u011Bk" . . "23520" . "Trojanov\u00E1, Jana" . "Czech Audio-Visual Speech Corpus for Recognition with Impaired Conditions"@en . "361914" . "S0284" . . . . . "Campr, Pavel" . "This is an audio-visual speech database for training and testing of Czech audio-visual continuous speech recognition systems collected with impaired illumination conditions. The corpus consists of about 20 hours of audio-visual records of 50 speakers in laboratory conditions. Recorded subjects were instructed to remain static. The illumination varied and chunks of each speaker were recorded with several different conditions, such as full illumination, or illumination from one side (left or right) only. These conditions make the database usable for training lip-/head-tracking systems under various illumination conditions independently of the language. Speakers were asked to read 200 sentences each (50 common for all speakers and 150 specific to each speaker). The average total length of recording per speaker was 23 minutes. Acoustic data are stored in wave files using PCM format, sampling frequency 44kHz, resolution 16 bits. Each speaker?s acoustic data set represents about 180 MB of disk space (" . "[B6D31A55DDD1]" . "RIV/49777513:23520/08:00503099!RIV10-AV0-23520___" . . . . . "This is an audio-visual speech database for training and testing of Czech audio-visual continuous speech recognition systems collected with impaired illumination conditions. The corpus consists of about 20 hours of audio-visual records of 50 speakers in laboratory conditions. Recorded subjects were instructed to remain static. The illumination varied and chunks of each speaker were recorded with several different conditions, such as full illumination, or illumination from one side (left or right) only. These conditions make the database usable for training lip-/head-tracking systems under various illumination conditions independently of the language. Speakers were asked to read 200 sentences each (50 common for all speakers and 150 specific to each speaker). The average total length of recording per speaker was 23 minutes. Acoustic data are stored in wave files using PCM format, sampling frequency 44kHz, resolution 16 bits. Each speaker?s acoustic data set represents about 180 MB of disk space ("@en . . . "European Language Resources Association (ELRA), http://catalog.elra.info/product_info.php?products_id=1082" . . "P(1ET101470416)" . "Czech Audio-Visual Speech Corpus for Recognition with Impaired Conditions" . "\u017Delezn\u00FD, Milo\u0161" . "SW produkt obsahuje speci\u00E1ln\u011B zpracovan\u00E9 z\u00E1znamy promluv v \u010Desk\u00E9m znakov\u00E9m jazyce pro pot\u0159eby tr\u00E9nov\u00E1n\u00ED syst\u00E9mu automatick\u00E9ho rozpozn\u00E1v\u00E1n\u00ED znak. \u0159e\u010Di. Pro nabyt\u00ED licence kontaktujte Valerii Mapell prost\u0159ednictv\u00EDm http://www.elda.org/contact.php?recip=vm" . . "RIV/49777513:23520/08:00503099" . "Hr\u00FAz, Marek" . .