Ontology type: schema:ScholarlyArticle
2017-05-31
AUTHORSTshering Dema, Margot Brereton, Jessica L. Cappadonna, Paul Roe, Anthony Truskinger, Jinglan Zhang
ABSTRACTMany ecologists are using acoustic monitoring to study animals and the health of ecosystems. Technological advances mean acoustic recording of nature can now be done at a relatively low cost, with minimal disturbance, and over long periods of time. Vast amounts of data are gathered yielding environmental soundscapes which requires new forms of visualization and interpretation of the data. Recently a novel visualization technique has been designed that represents soundscapes using dense visual summaries of acoustic patterns. However, little is known about how this visualization tool can be employed to make sense of soundscapes. Understanding how the technique can be best used and developed requires collaboration between interface, algorithm designers and ecologists. We empirically investigated the practices and needs of ecologists using acoustic monitoring technologies. In particular, we investigated the use of the soundscape visualization tool by teams of ecologists researching endangered species detection, species behaviour, and monitoring of ecological areas using long duration audio recordings. Our findings highlight the opportunities and challenges that ecologists face in making sense of large acoustic datasets through patterns of acoustic events. We reveal the characteristic processes for collaboratively generating situated accounts of natural places from soundscapes using visualization. We also discuss the biases inherent in the approach. Big data from nature has different characteristics from social and informational data sources that comprise much of the World Wide Web. We conclude with design implications for visual interfaces to facilitate collaborative exploration and discovery through soundscapes. More... »
PAGES693-731
http://scigraph.springernature.com/pub.10.1007/s10606-017-9286-9
DOIhttp://dx.doi.org/10.1007/s10606-017-9286-9
DIMENSIONShttps://app.dimensions.ai/details/publication/pub.1085730245
JSON-LD is the canonical representation for SciGraph data.
TIP: You can open this SciGraph record using an external JSON-LD service: JSON-LD Playground Google SDTT
[
{
"@context": "https://springernature.github.io/scigraph/jsonld/sgcontext.json",
"about": [
{
"id": "http://purl.org/au-research/vocabulary/anzsrc-for/2008/08",
"inDefinedTermSet": "http://purl.org/au-research/vocabulary/anzsrc-for/2008/",
"name": "Information and Computing Sciences",
"type": "DefinedTerm"
},
{
"id": "http://purl.org/au-research/vocabulary/anzsrc-for/2008/0801",
"inDefinedTermSet": "http://purl.org/au-research/vocabulary/anzsrc-for/2008/",
"name": "Artificial Intelligence and Image Processing",
"type": "DefinedTerm"
}
],
"author": [
{
"affiliation": {
"alternateName": "Computer Human Interaction, Science and Engineering Faculty, Queensland University of Technology, 2 George St, QLD 4000,, Brisbane, Australia",
"id": "http://www.grid.ac/institutes/grid.1024.7",
"name": [
"Computer Human Interaction, Science and Engineering Faculty, Queensland University of Technology, 2 George St, QLD 4000,, Brisbane, Australia"
],
"type": "Organization"
},
"familyName": "Dema",
"givenName": "Tshering",
"id": "sg:person.013050115036.93",
"sameAs": [
"https://app.dimensions.ai/discover/publication?and_facet_researcher=ur.013050115036.93"
],
"type": "Person"
},
{
"affiliation": {
"alternateName": "Computer Human Interaction, Science and Engineering Faculty, Queensland University of Technology, 2 George St, QLD 4000,, Brisbane, Australia",
"id": "http://www.grid.ac/institutes/grid.1024.7",
"name": [
"Computer Human Interaction, Science and Engineering Faculty, Queensland University of Technology, 2 George St, QLD 4000,, Brisbane, Australia"
],
"type": "Organization"
},
"familyName": "Brereton",
"givenName": "Margot",
"id": "sg:person.015764524414.17",
"sameAs": [
"https://app.dimensions.ai/discover/publication?and_facet_researcher=ur.015764524414.17"
],
"type": "Person"
},
{
"affiliation": {
"alternateName": "Computer Human Interaction, Science and Engineering Faculty, Queensland University of Technology, 2 George St, QLD 4000,, Brisbane, Australia",
"id": "http://www.grid.ac/institutes/grid.1024.7",
"name": [
"Computer Human Interaction, Science and Engineering Faculty, Queensland University of Technology, 2 George St, QLD 4000,, Brisbane, Australia"
],
"type": "Organization"
},
"familyName": "Cappadonna",
"givenName": "Jessica L.",
"id": "sg:person.014443056036.73",
"sameAs": [
"https://app.dimensions.ai/discover/publication?and_facet_researcher=ur.014443056036.73"
],
"type": "Person"
},
{
"affiliation": {
"alternateName": "Computer Human Interaction, Science and Engineering Faculty, Queensland University of Technology, 2 George St, QLD 4000,, Brisbane, Australia",
"id": "http://www.grid.ac/institutes/grid.1024.7",
"name": [
"Computer Human Interaction, Science and Engineering Faculty, Queensland University of Technology, 2 George St, QLD 4000,, Brisbane, Australia"
],
"type": "Organization"
},
"familyName": "Roe",
"givenName": "Paul",
"id": "sg:person.01147472725.88",
"sameAs": [
"https://app.dimensions.ai/discover/publication?and_facet_researcher=ur.01147472725.88"
],
"type": "Person"
},
{
"affiliation": {
"alternateName": "Computer Human Interaction, Science and Engineering Faculty, Queensland University of Technology, 2 George St, QLD 4000,, Brisbane, Australia",
"id": "http://www.grid.ac/institutes/grid.1024.7",
"name": [
"Computer Human Interaction, Science and Engineering Faculty, Queensland University of Technology, 2 George St, QLD 4000,, Brisbane, Australia"
],
"type": "Organization"
},
"familyName": "Truskinger",
"givenName": "Anthony",
"id": "sg:person.012070356157.61",
"sameAs": [
"https://app.dimensions.ai/discover/publication?and_facet_researcher=ur.012070356157.61"
],
"type": "Person"
},
{
"affiliation": {
"alternateName": "Computer Human Interaction, Science and Engineering Faculty, Queensland University of Technology, 2 George St, QLD 4000,, Brisbane, Australia",
"id": "http://www.grid.ac/institutes/grid.1024.7",
"name": [
"Computer Human Interaction, Science and Engineering Faculty, Queensland University of Technology, 2 George St, QLD 4000,, Brisbane, Australia"
],
"type": "Organization"
},
"familyName": "Zhang",
"givenName": "Jinglan",
"id": "sg:person.015240111037.43",
"sameAs": [
"https://app.dimensions.ai/discover/publication?and_facet_researcher=ur.015240111037.43"
],
"type": "Person"
}
],
"citation": [
{
"id": "sg:pub.10.1038/ncomms12558",
"sameAs": [
"https://app.dimensions.ai/details/publication/pub.1022030358",
"https://doi.org/10.1038/ncomms12558"
],
"type": "CreativeWork"
},
{
"id": "sg:pub.10.1038/nature14324",
"sameAs": [
"https://app.dimensions.ai/details/publication/pub.1032466158",
"https://doi.org/10.1038/nature14324"
],
"type": "CreativeWork"
},
{
"id": "sg:pub.10.1007/978-3-319-02612-1_3",
"sameAs": [
"https://app.dimensions.ai/details/publication/pub.1014865847",
"https://doi.org/10.1007/978-3-319-02612-1_3"
],
"type": "CreativeWork"
},
{
"id": "sg:pub.10.1007/978-3-319-20499-4_17",
"sameAs": [
"https://app.dimensions.ai/details/publication/pub.1050551603",
"https://doi.org/10.1007/978-3-319-20499-4_17"
],
"type": "CreativeWork"
},
{
"id": "sg:pub.10.1007/978-3-642-36973-5_3",
"sameAs": [
"https://app.dimensions.ai/details/publication/pub.1022104390",
"https://doi.org/10.1007/978-3-642-36973-5_3"
],
"type": "CreativeWork"
}
],
"datePublished": "2017-05-31",
"datePublishedReg": "2017-05-31",
"description": "Many ecologists are using acoustic monitoring to study animals and the health of ecosystems. Technological advances mean acoustic recording of nature can now be done at a relatively low cost, with minimal disturbance, and over long periods of time. Vast amounts of data are gathered yielding environmental soundscapes which requires new forms of visualization and interpretation of the data. Recently a novel visualization technique has been designed that represents soundscapes using dense visual summaries of acoustic patterns. However, little is known about how this visualization tool can be employed to make sense of soundscapes. Understanding how the technique can be best used and developed requires collaboration between interface, algorithm designers and ecologists. We empirically investigated the practices and needs of ecologists using acoustic monitoring technologies. In particular, we investigated the use of the soundscape visualization tool by teams of ecologists researching endangered species detection, species behaviour, and monitoring of ecological areas using long duration audio recordings. Our findings highlight the opportunities and challenges that ecologists face in making sense of large acoustic datasets through patterns of acoustic events. We reveal the characteristic processes for collaboratively generating situated accounts of natural places from soundscapes using visualization. We also discuss the biases inherent in the approach. Big data from nature has different characteristics from social and informational data sources that comprise much of the World Wide Web. We conclude with design implications for visual interfaces to facilitate collaborative exploration and discovery through soundscapes.",
"genre": "article",
"id": "sg:pub.10.1007/s10606-017-9286-9",
"isAccessibleForFree": false,
"isPartOf": [
{
"id": "sg:journal.1356904",
"issn": [
"1431-1496",
"1573-7551"
],
"name": "Computer Supported Cooperative Work (CSCW)",
"publisher": "Springer Nature",
"type": "Periodical"
},
{
"issueNumber": "4-6",
"type": "PublicationIssue"
},
{
"type": "PublicationVolume",
"volumeNumber": "26"
}
],
"keywords": [
"visualization tools",
"long-duration audio recordings",
"collaborative exploration",
"World Wide Web",
"environmental sound data",
"novel visualization technique",
"algorithm designers",
"large acoustic datasets",
"big data",
"visual interface",
"teams of ecologists",
"Wide Web",
"visualization techniques",
"visual summary",
"needs of ecologists",
"vast amount",
"acoustic datasets",
"acoustic events",
"data sources",
"design implications",
"acoustic monitoring technology",
"monitoring technology",
"audio recordings",
"sound data",
"visualization",
"acoustic patterns",
"different characteristics",
"technological advances",
"low cost",
"tool",
"interface",
"dataset",
"environmental soundscapes",
"designers",
"Web",
"acoustic recordings",
"acoustic monitoring",
"data",
"soundscape",
"new forms",
"technique",
"technology",
"monitoring",
"exploration",
"collaboration",
"sensemaking",
"cost",
"detection",
"challenges",
"team",
"need",
"discovery",
"advances",
"sense",
"species detection",
"natural place",
"recordings",
"opportunities",
"time",
"use",
"process",
"nature",
"amount",
"patterns",
"area",
"account",
"source",
"minimal disturbance",
"practice",
"characteristics",
"ecologists",
"ecosystems",
"interpretation",
"behavior",
"place",
"biases",
"long period",
"form",
"summary",
"events",
"health of ecosystems",
"species behavior",
"disturbances",
"characteristic processes",
"implications",
"health",
"ecological areas",
"findings",
"period",
"animals",
"approach"
],
"name": "Collaborative Exploration and Sensemaking of Big Environmental Sound Data",
"pagination": "693-731",
"productId": [
{
"name": "dimensions_id",
"type": "PropertyValue",
"value": [
"pub.1085730245"
]
},
{
"name": "doi",
"type": "PropertyValue",
"value": [
"10.1007/s10606-017-9286-9"
]
}
],
"sameAs": [
"https://doi.org/10.1007/s10606-017-9286-9",
"https://app.dimensions.ai/details/publication/pub.1085730245"
],
"sdDataset": "articles",
"sdDatePublished": "2022-08-04T17:06",
"sdLicense": "https://scigraph.springernature.com/explorer/license/",
"sdPublisher": {
"name": "Springer Nature - SN SciGraph project",
"type": "Organization"
},
"sdSource": "s3://com-springernature-scigraph/baseset/20220804/entities/gbq_results/article/article_750.jsonl",
"type": "ScholarlyArticle",
"url": "https://doi.org/10.1007/s10606-017-9286-9"
}
]
Download the RDF metadata as: json-ld nt turtle xml License info
JSON-LD is a popular format for linked data which is fully compatible with JSON.
curl -H 'Accept: application/ld+json' 'https://scigraph.springernature.com/pub.10.1007/s10606-017-9286-9'
N-Triples is a line-based linked data format ideal for batch operations.
curl -H 'Accept: application/n-triples' 'https://scigraph.springernature.com/pub.10.1007/s10606-017-9286-9'
Turtle is a human-readable linked data format.
curl -H 'Accept: text/turtle' 'https://scigraph.springernature.com/pub.10.1007/s10606-017-9286-9'
RDF/XML is a standard XML format for linked data.
curl -H 'Accept: application/rdf+xml' 'https://scigraph.springernature.com/pub.10.1007/s10606-017-9286-9'
This table displays all metadata directly associated to this object as RDF triples.
203 TRIPLES
21 PREDICATES
120 URIs
107 LITERALS
6 BLANK NODES
Subject | Predicate | Object | |
---|---|---|---|
1 | sg:pub.10.1007/s10606-017-9286-9 | schema:about | anzsrc-for:08 |
2 | ″ | ″ | anzsrc-for:0801 |
3 | ″ | schema:author | N7dcb57dff884478ca67c246d5bb4774c |
4 | ″ | schema:citation | sg:pub.10.1007/978-3-319-02612-1_3 |
5 | ″ | ″ | sg:pub.10.1007/978-3-319-20499-4_17 |
6 | ″ | ″ | sg:pub.10.1007/978-3-642-36973-5_3 |
7 | ″ | ″ | sg:pub.10.1038/nature14324 |
8 | ″ | ″ | sg:pub.10.1038/ncomms12558 |
9 | ″ | schema:datePublished | 2017-05-31 |
10 | ″ | schema:datePublishedReg | 2017-05-31 |
11 | ″ | schema:description | Many ecologists are using acoustic monitoring to study animals and the health of ecosystems. Technological advances mean acoustic recording of nature can now be done at a relatively low cost, with minimal disturbance, and over long periods of time. Vast amounts of data are gathered yielding environmental soundscapes which requires new forms of visualization and interpretation of the data. Recently a novel visualization technique has been designed that represents soundscapes using dense visual summaries of acoustic patterns. However, little is known about how this visualization tool can be employed to make sense of soundscapes. Understanding how the technique can be best used and developed requires collaboration between interface, algorithm designers and ecologists. We empirically investigated the practices and needs of ecologists using acoustic monitoring technologies. In particular, we investigated the use of the soundscape visualization tool by teams of ecologists researching endangered species detection, species behaviour, and monitoring of ecological areas using long duration audio recordings. Our findings highlight the opportunities and challenges that ecologists face in making sense of large acoustic datasets through patterns of acoustic events. We reveal the characteristic processes for collaboratively generating situated accounts of natural places from soundscapes using visualization. We also discuss the biases inherent in the approach. Big data from nature has different characteristics from social and informational data sources that comprise much of the World Wide Web. We conclude with design implications for visual interfaces to facilitate collaborative exploration and discovery through soundscapes. |
12 | ″ | schema:genre | article |
13 | ″ | schema:isAccessibleForFree | false |
14 | ″ | schema:isPartOf | N1275d692345f4977bdc6f82c51efdf1d |
15 | ″ | ″ | N4176acf1ee514ad29a9c5b43fcb5a0ee |
16 | ″ | ″ | sg:journal.1356904 |
17 | ″ | schema:keywords | Web |
18 | ″ | ″ | Wide Web |
19 | ″ | ″ | World Wide Web |
20 | ″ | ″ | account |
21 | ″ | ″ | acoustic datasets |
22 | ″ | ″ | acoustic events |
23 | ″ | ″ | acoustic monitoring |
24 | ″ | ″ | acoustic monitoring technology |
25 | ″ | ″ | acoustic patterns |
26 | ″ | ″ | acoustic recordings |
27 | ″ | ″ | advances |
28 | ″ | ″ | algorithm designers |
29 | ″ | ″ | amount |
30 | ″ | ″ | animals |
31 | ″ | ″ | approach |
32 | ″ | ″ | area |
33 | ″ | ″ | audio recordings |
34 | ″ | ″ | behavior |
35 | ″ | ″ | biases |
36 | ″ | ″ | big data |
37 | ″ | ″ | challenges |
38 | ″ | ″ | characteristic processes |
39 | ″ | ″ | characteristics |
40 | ″ | ″ | collaboration |
41 | ″ | ″ | collaborative exploration |
42 | ″ | ″ | cost |
43 | ″ | ″ | data |
44 | ″ | ″ | data sources |
45 | ″ | ″ | dataset |
46 | ″ | ″ | design implications |
47 | ″ | ″ | designers |
48 | ″ | ″ | detection |
49 | ″ | ″ | different characteristics |
50 | ″ | ″ | discovery |
51 | ″ | ″ | disturbances |
52 | ″ | ″ | ecological areas |
53 | ″ | ″ | ecologists |
54 | ″ | ″ | ecosystems |
55 | ″ | ″ | environmental sound data |
56 | ″ | ″ | environmental soundscapes |
57 | ″ | ″ | events |
58 | ″ | ″ | exploration |
59 | ″ | ″ | findings |
60 | ″ | ″ | form |
61 | ″ | ″ | health |
62 | ″ | ″ | health of ecosystems |
63 | ″ | ″ | implications |
64 | ″ | ″ | interface |
65 | ″ | ″ | interpretation |
66 | ″ | ″ | large acoustic datasets |
67 | ″ | ″ | long period |
68 | ″ | ″ | long-duration audio recordings |
69 | ″ | ″ | low cost |
70 | ″ | ″ | minimal disturbance |
71 | ″ | ″ | monitoring |
72 | ″ | ″ | monitoring technology |
73 | ″ | ″ | natural place |
74 | ″ | ″ | nature |
75 | ″ | ″ | need |
76 | ″ | ″ | needs of ecologists |
77 | ″ | ″ | new forms |
78 | ″ | ″ | novel visualization technique |
79 | ″ | ″ | opportunities |
80 | ″ | ″ | patterns |
81 | ″ | ″ | period |
82 | ″ | ″ | place |
83 | ″ | ″ | practice |
84 | ″ | ″ | process |
85 | ″ | ″ | recordings |
86 | ″ | ″ | sense |
87 | ″ | ″ | sensemaking |
88 | ″ | ″ | sound data |
89 | ″ | ″ | soundscape |
90 | ″ | ″ | source |
91 | ″ | ″ | species behavior |
92 | ″ | ″ | species detection |
93 | ″ | ″ | summary |
94 | ″ | ″ | team |
95 | ″ | ″ | teams of ecologists |
96 | ″ | ″ | technique |
97 | ″ | ″ | technological advances |
98 | ″ | ″ | technology |
99 | ″ | ″ | time |
100 | ″ | ″ | tool |
101 | ″ | ″ | use |
102 | ″ | ″ | vast amount |
103 | ″ | ″ | visual interface |
104 | ″ | ″ | visual summary |
105 | ″ | ″ | visualization |
106 | ″ | ″ | visualization techniques |
107 | ″ | ″ | visualization tools |
108 | ″ | schema:name | Collaborative Exploration and Sensemaking of Big Environmental Sound Data |
109 | ″ | schema:pagination | 693-731 |
110 | ″ | schema:productId | N0761b62ec39849448d14068940b78abd |
111 | ″ | ″ | Nd78a71e46e5645a9879de5adb19add09 |
112 | ″ | schema:sameAs | https://app.dimensions.ai/details/publication/pub.1085730245 |
113 | ″ | ″ | https://doi.org/10.1007/s10606-017-9286-9 |
114 | ″ | schema:sdDatePublished | 2022-08-04T17:06 |
115 | ″ | schema:sdLicense | https://scigraph.springernature.com/explorer/license/ |
116 | ″ | schema:sdPublisher | Neabc319f94ed4b45ac59a9d99e2338ce |
117 | ″ | schema:url | https://doi.org/10.1007/s10606-017-9286-9 |
118 | ″ | sgo:license | sg:explorer/license/ |
119 | ″ | sgo:sdDataset | articles |
120 | ″ | rdf:type | schema:ScholarlyArticle |
121 | N0761b62ec39849448d14068940b78abd | schema:name | dimensions_id |
122 | ″ | schema:value | pub.1085730245 |
123 | ″ | rdf:type | schema:PropertyValue |
124 | N1275d692345f4977bdc6f82c51efdf1d | schema:volumeNumber | 26 |
125 | ″ | rdf:type | schema:PublicationVolume |
126 | N13eb09a7877a4889a608eddcdfc0ef7f | rdf:first | sg:person.012070356157.61 |
127 | ″ | rdf:rest | N3800d95fb5964a7e9dcf96f6af6137a2 |
128 | N1c4aa7e97bca4fcf9ed36b828689b460 | rdf:first | sg:person.01147472725.88 |
129 | ″ | rdf:rest | N13eb09a7877a4889a608eddcdfc0ef7f |
130 | N3800d95fb5964a7e9dcf96f6af6137a2 | rdf:first | sg:person.015240111037.43 |
131 | ″ | rdf:rest | rdf:nil |
132 | N4176acf1ee514ad29a9c5b43fcb5a0ee | schema:issueNumber | 4-6 |
133 | ″ | rdf:type | schema:PublicationIssue |
134 | N7dcb57dff884478ca67c246d5bb4774c | rdf:first | sg:person.013050115036.93 |
135 | ″ | rdf:rest | N7e08a57d73194cabac732aad27c7e7fa |
136 | N7e08a57d73194cabac732aad27c7e7fa | rdf:first | sg:person.015764524414.17 |
137 | ″ | rdf:rest | Nc1d1c05d8b1649a681c7707e92dbcef7 |
138 | Nc1d1c05d8b1649a681c7707e92dbcef7 | rdf:first | sg:person.014443056036.73 |
139 | ″ | rdf:rest | N1c4aa7e97bca4fcf9ed36b828689b460 |
140 | Nd78a71e46e5645a9879de5adb19add09 | schema:name | doi |
141 | ″ | schema:value | 10.1007/s10606-017-9286-9 |
142 | ″ | rdf:type | schema:PropertyValue |
143 | Neabc319f94ed4b45ac59a9d99e2338ce | schema:name | Springer Nature - SN SciGraph project |
144 | ″ | rdf:type | schema:Organization |
145 | anzsrc-for:08 | schema:inDefinedTermSet | anzsrc-for: |
146 | ″ | schema:name | Information and Computing Sciences |
147 | ″ | rdf:type | schema:DefinedTerm |
148 | anzsrc-for:0801 | schema:inDefinedTermSet | anzsrc-for: |
149 | ″ | schema:name | Artificial Intelligence and Image Processing |
150 | ″ | rdf:type | schema:DefinedTerm |
151 | sg:journal.1356904 | schema:issn | 1431-1496 |
152 | ″ | ″ | 1573-7551 |
153 | ″ | schema:name | Computer Supported Cooperative Work (CSCW) |
154 | ″ | schema:publisher | Springer Nature |
155 | ″ | rdf:type | schema:Periodical |
156 | sg:person.01147472725.88 | schema:affiliation | grid-institutes:grid.1024.7 |
157 | ″ | schema:familyName | Roe |
158 | ″ | schema:givenName | Paul |
159 | ″ | schema:sameAs | https://app.dimensions.ai/discover/publication?and_facet_researcher=ur.01147472725.88 |
160 | ″ | rdf:type | schema:Person |
161 | sg:person.012070356157.61 | schema:affiliation | grid-institutes:grid.1024.7 |
162 | ″ | schema:familyName | Truskinger |
163 | ″ | schema:givenName | Anthony |
164 | ″ | schema:sameAs | https://app.dimensions.ai/discover/publication?and_facet_researcher=ur.012070356157.61 |
165 | ″ | rdf:type | schema:Person |
166 | sg:person.013050115036.93 | schema:affiliation | grid-institutes:grid.1024.7 |
167 | ″ | schema:familyName | Dema |
168 | ″ | schema:givenName | Tshering |
169 | ″ | schema:sameAs | https://app.dimensions.ai/discover/publication?and_facet_researcher=ur.013050115036.93 |
170 | ″ | rdf:type | schema:Person |
171 | sg:person.014443056036.73 | schema:affiliation | grid-institutes:grid.1024.7 |
172 | ″ | schema:familyName | Cappadonna |
173 | ″ | schema:givenName | Jessica L. |
174 | ″ | schema:sameAs | https://app.dimensions.ai/discover/publication?and_facet_researcher=ur.014443056036.73 |
175 | ″ | rdf:type | schema:Person |
176 | sg:person.015240111037.43 | schema:affiliation | grid-institutes:grid.1024.7 |
177 | ″ | schema:familyName | Zhang |
178 | ″ | schema:givenName | Jinglan |
179 | ″ | schema:sameAs | https://app.dimensions.ai/discover/publication?and_facet_researcher=ur.015240111037.43 |
180 | ″ | rdf:type | schema:Person |
181 | sg:person.015764524414.17 | schema:affiliation | grid-institutes:grid.1024.7 |
182 | ″ | schema:familyName | Brereton |
183 | ″ | schema:givenName | Margot |
184 | ″ | schema:sameAs | https://app.dimensions.ai/discover/publication?and_facet_researcher=ur.015764524414.17 |
185 | ″ | rdf:type | schema:Person |
186 | sg:pub.10.1007/978-3-319-02612-1_3 | schema:sameAs | https://app.dimensions.ai/details/publication/pub.1014865847 |
187 | ″ | ″ | https://doi.org/10.1007/978-3-319-02612-1_3 |
188 | ″ | rdf:type | schema:CreativeWork |
189 | sg:pub.10.1007/978-3-319-20499-4_17 | schema:sameAs | https://app.dimensions.ai/details/publication/pub.1050551603 |
190 | ″ | ″ | https://doi.org/10.1007/978-3-319-20499-4_17 |
191 | ″ | rdf:type | schema:CreativeWork |
192 | sg:pub.10.1007/978-3-642-36973-5_3 | schema:sameAs | https://app.dimensions.ai/details/publication/pub.1022104390 |
193 | ″ | ″ | https://doi.org/10.1007/978-3-642-36973-5_3 |
194 | ″ | rdf:type | schema:CreativeWork |
195 | sg:pub.10.1038/nature14324 | schema:sameAs | https://app.dimensions.ai/details/publication/pub.1032466158 |
196 | ″ | ″ | https://doi.org/10.1038/nature14324 |
197 | ″ | rdf:type | schema:CreativeWork |
198 | sg:pub.10.1038/ncomms12558 | schema:sameAs | https://app.dimensions.ai/details/publication/pub.1022030358 |
199 | ″ | ″ | https://doi.org/10.1038/ncomms12558 |
200 | ″ | rdf:type | schema:CreativeWork |
201 | grid-institutes:grid.1024.7 | schema:alternateName | Computer Human Interaction, Science and Engineering Faculty, Queensland University of Technology, 2 George St, QLD 4000,, Brisbane, Australia |
202 | ″ | schema:name | Computer Human Interaction, Science and Engineering Faculty, Queensland University of Technology, 2 George St, QLD 4000,, Brisbane, Australia |
203 | ″ | rdf:type | schema:Organization |