Ontology type: schema:Chapter Open Access: True
2014
AUTHORSHelen L. Bear , Richard W. Harvey , Barry-John Theobald , Yuxuan Lan
ABSTRACTA critical assumption of all current visual speech recognition systems is that there are visual speech units called visemes which can be mapped to units of acoustic speech, the phonemes. Despite there being a number of published maps it is infrequent to see the effectiveness of these tested, particularly on visual-only lip-reading (many works use audio-visual speech). Here we examine 120 mappings and consider if any are stable across talkers. We show a method for devising maps based on phoneme confusions from an automated lip-reading system, and we present new mappings that show improvements for individual talkers. More... »
PAGES230-239
Advances in Visual Computing
ISBN
978-3-319-14363-7
978-3-319-14364-4
http://scigraph.springernature.com/pub.10.1007/978-3-319-14364-4_22
DOIhttp://dx.doi.org/10.1007/978-3-319-14364-4_22
DIMENSIONShttps://app.dimensions.ai/details/publication/pub.1044482572
JSON-LD is the canonical representation for SciGraph data.
TIP: You can open this SciGraph record using an external JSON-LD service: JSON-LD Playground Google SDTT
[
{
"@context": "https://springernature.github.io/scigraph/jsonld/sgcontext.json",
"about": [
{
"id": "http://purl.org/au-research/vocabulary/anzsrc-for/2008/0801",
"inDefinedTermSet": "http://purl.org/au-research/vocabulary/anzsrc-for/2008/",
"name": "Artificial Intelligence and Image Processing",
"type": "DefinedTerm"
},
{
"id": "http://purl.org/au-research/vocabulary/anzsrc-for/2008/08",
"inDefinedTermSet": "http://purl.org/au-research/vocabulary/anzsrc-for/2008/",
"name": "Information and Computing Sciences",
"type": "DefinedTerm"
}
],
"author": [
{
"affiliation": {
"alternateName": "University of East Anglia",
"id": "https://www.grid.ac/institutes/grid.8273.e",
"name": [
"School of Computing Sciences, University of East Anglia, Norwich, UK"
],
"type": "Organization"
},
"familyName": "Bear",
"givenName": "Helen L.",
"id": "sg:person.010310203152.09",
"sameAs": [
"https://app.dimensions.ai/discover/publication?and_facet_researcher=ur.010310203152.09"
],
"type": "Person"
},
{
"affiliation": {
"alternateName": "University of East Anglia",
"id": "https://www.grid.ac/institutes/grid.8273.e",
"name": [
"School of Computing Sciences, University of East Anglia, Norwich, UK"
],
"type": "Organization"
},
"familyName": "Harvey",
"givenName": "Richard W.",
"id": "sg:person.013620351211.30",
"sameAs": [
"https://app.dimensions.ai/discover/publication?and_facet_researcher=ur.013620351211.30"
],
"type": "Person"
},
{
"affiliation": {
"alternateName": "University of East Anglia",
"id": "https://www.grid.ac/institutes/grid.8273.e",
"name": [
"School of Computing Sciences, University of East Anglia, Norwich, UK"
],
"type": "Organization"
},
"familyName": "Theobald",
"givenName": "Barry-John",
"id": "sg:person.010544173235.71",
"sameAs": [
"https://app.dimensions.ai/discover/publication?and_facet_researcher=ur.010544173235.71"
],
"type": "Person"
},
{
"affiliation": {
"alternateName": "University of East Anglia",
"id": "https://www.grid.ac/institutes/grid.8273.e",
"name": [
"School of Computing Sciences, University of East Anglia, Norwich, UK"
],
"type": "Organization"
},
"familyName": "Lan",
"givenName": "Yuxuan",
"id": "sg:person.016265032243.58",
"sameAs": [
"https://app.dimensions.ai/discover/publication?and_facet_researcher=ur.016265032243.58"
],
"type": "Person"
}
],
"citation": [
{
"id": "https://doi.org/10.1044/jshd.4104.530",
"sameAs": [
"https://app.dimensions.ai/details/publication/pub.1005211937"
],
"type": "CreativeWork"
},
{
"id": "https://doi.org/10.1044/jshr.1104.796",
"sameAs": [
"https://app.dimensions.ai/details/publication/pub.1014439870"
],
"type": "CreativeWork"
},
{
"id": "https://doi.org/10.1044/jshr.2803.381",
"sameAs": [
"https://app.dimensions.ai/details/publication/pub.1016068875"
],
"type": "CreativeWork"
},
{
"id": "sg:pub.10.1023/b:visi.0000029666.37597.d3",
"sameAs": [
"https://app.dimensions.ai/details/publication/pub.1022700327",
"https://doi.org/10.1023/b:visi.0000029666.37597.d3"
],
"type": "CreativeWork"
},
{
"id": "https://doi.org/10.1044/jshr.2001.130",
"sameAs": [
"https://app.dimensions.ai/details/publication/pub.1027835888"
],
"type": "CreativeWork"
},
{
"id": "sg:pub.10.1007/3-540-45683-x_60",
"sameAs": [
"https://app.dimensions.ai/details/publication/pub.1038672253",
"https://doi.org/10.1007/3-540-45683-x_60"
],
"type": "CreativeWork"
},
{
"id": "https://doi.org/10.1016/0167-8655(88)90094-3",
"sameAs": [
"https://app.dimensions.ai/details/publication/pub.1040847185"
],
"type": "CreativeWork"
},
{
"id": "https://doi.org/10.1016/0167-8655(88)90094-3",
"sameAs": [
"https://app.dimensions.ai/details/publication/pub.1040847185"
],
"type": "CreativeWork"
},
{
"id": "https://doi.org/10.1044/jshr.1503.474",
"sameAs": [
"https://app.dimensions.ai/details/publication/pub.1046023900"
],
"type": "CreativeWork"
},
{
"id": "https://doi.org/10.1044/jshr.0303.212",
"sameAs": [
"https://app.dimensions.ai/details/publication/pub.1049929020"
],
"type": "CreativeWork"
},
{
"id": "https://doi.org/10.1109/5.664274",
"sameAs": [
"https://app.dimensions.ai/details/publication/pub.1061179889"
],
"type": "CreativeWork"
},
{
"id": "https://doi.org/10.1121/1.389537",
"sameAs": [
"https://app.dimensions.ai/details/publication/pub.1062340829"
],
"type": "CreativeWork"
}
],
"datePublished": "2014",
"datePublishedReg": "2014-01-01",
"description": "A critical assumption of all current visual speech recognition systems is that there are visual speech units called visemes which can be mapped to units of acoustic speech, the phonemes. Despite there being a number of published maps it is infrequent to see the effectiveness of these tested, particularly on visual-only lip-reading (many works use audio-visual speech). Here we examine 120 mappings and consider if any are stable across talkers. We show a method for devising maps based on phoneme confusions from an automated lip-reading system, and we present new mappings that show improvements for individual talkers.",
"editor": [
{
"familyName": "Bebis",
"givenName": "George",
"type": "Person"
},
{
"familyName": "Boyle",
"givenName": "Richard",
"type": "Person"
},
{
"familyName": "Parvin",
"givenName": "Bahram",
"type": "Person"
},
{
"familyName": "Koracin",
"givenName": "Darko",
"type": "Person"
},
{
"familyName": "McMahan",
"givenName": "Ryan",
"type": "Person"
},
{
"familyName": "Jerald",
"givenName": "Jason",
"type": "Person"
},
{
"familyName": "Zhang",
"givenName": "Hui",
"type": "Person"
},
{
"familyName": "Drucker",
"givenName": "Steven M.",
"type": "Person"
},
{
"familyName": "Kambhamettu",
"givenName": "Chandra",
"type": "Person"
},
{
"familyName": "El Choubassi",
"givenName": "Maha",
"type": "Person"
},
{
"familyName": "Deng",
"givenName": "Zhigang",
"type": "Person"
},
{
"familyName": "Carlson",
"givenName": "Mark",
"type": "Person"
}
],
"genre": "chapter",
"id": "sg:pub.10.1007/978-3-319-14364-4_22",
"inLanguage": [
"en"
],
"isAccessibleForFree": true,
"isPartOf": {
"isbn": [
"978-3-319-14363-7",
"978-3-319-14364-4"
],
"name": "Advances in Visual Computing",
"type": "Book"
},
"name": "Which Phoneme-to-Viseme Maps Best Improve Visual-Only Computer Lip-Reading?",
"pagination": "230-239",
"productId": [
{
"name": "doi",
"type": "PropertyValue",
"value": [
"10.1007/978-3-319-14364-4_22"
]
},
{
"name": "readcube_id",
"type": "PropertyValue",
"value": [
"d3e31fefc562f1bb86ccc4b19c8d3b8d9ad1619b07aa0cc66a239e39628f8431"
]
},
{
"name": "dimensions_id",
"type": "PropertyValue",
"value": [
"pub.1044482572"
]
}
],
"publisher": {
"location": "Cham",
"name": "Springer International Publishing",
"type": "Organisation"
},
"sameAs": [
"https://doi.org/10.1007/978-3-319-14364-4_22",
"https://app.dimensions.ai/details/publication/pub.1044482572"
],
"sdDataset": "chapters",
"sdDatePublished": "2019-04-15T13:30",
"sdLicense": "https://scigraph.springernature.com/explorer/license/",
"sdPublisher": {
"name": "Springer Nature - SN SciGraph project",
"type": "Organization"
},
"sdSource": "s3://com-uberresearch-data-dimensions-target-20181106-alternative/cleanup/v134/2549eaecd7973599484d7c17b260dba0a4ecb94b/merge/v9/a6c9fde33151104705d4d7ff012ea9563521a3ce/jats-lookup/v90/0000000001_0000000264/records_8664_00000270.jsonl",
"type": "Chapter",
"url": "http://link.springer.com/10.1007/978-3-319-14364-4_22"
}
]
Download the RDF metadata as: json-ld nt turtle xml License info
JSON-LD is a popular format for linked data which is fully compatible with JSON.
curl -H 'Accept: application/ld+json' 'https://scigraph.springernature.com/pub.10.1007/978-3-319-14364-4_22'
N-Triples is a line-based linked data format ideal for batch operations.
curl -H 'Accept: application/n-triples' 'https://scigraph.springernature.com/pub.10.1007/978-3-319-14364-4_22'
Turtle is a human-readable linked data format.
curl -H 'Accept: text/turtle' 'https://scigraph.springernature.com/pub.10.1007/978-3-319-14364-4_22'
RDF/XML is a standard XML format for linked data.
curl -H 'Accept: application/rdf+xml' 'https://scigraph.springernature.com/pub.10.1007/978-3-319-14364-4_22'
This table displays all metadata directly associated to this object as RDF triples.
176 TRIPLES
23 PREDICATES
38 URIs
20 LITERALS
8 BLANK NODES