ORCID Profile
0000-0002-6387-4181
Does something not look right? The information on this page has been harvested from data sources that may not be up to date. We continue to work with information providers to improve coverage and quality. To report an issue, use the Feedback Form.
In Research Link Australia (RLA), "Research Topics" refer to ANZSRC FOR and SEO codes. These topics are either sourced from ANZSRC FOR and SEO codes listed in researchers' related grants or generated by a large language model (LLM) based on their publications.
Psychology | Sensory Processes, Perception And Performance | Learning, Memory, Cognition And Language | Sensory Processes, Perception and Performance | Computational Linguistics | Computer Perception, Memory And Attention | Developmental Psychology and Ageing | Computer-Human Interaction | Biological Psychology (Neuropsychology, Psychopharmacology, | Neurocognitive Patterns And Neural Networks | Laboratory Phonetics And Speech Science | Linguistic Processes (incl. Speech Production and Comprehension) | Speech Recognition
Behavioural and cognitive sciences | Expanding Knowledge in Psychology and Cognitive Sciences | Information processing services | Communication services not elsewhere classified | Hearing, vision, speech and their disorders | Health Related to Ageing | Biological sciences |
Publisher: Informa UK Limited
Date: 09-2012
Publisher: Elsevier BV
Date: 10-2014
DOI: 10.1016/J.BANDL.2014.07.012
Abstract: Auditory speech processing is facilitated when the talker's face/head movements are seen. This effect is typically explained in terms of visual speech providing form and/or timing information. We determined the effect of both types of information on a speech/non-speech task (non-speech stimuli were spectrally rotated speech). All stimuli were presented paired with the talker's static or moving face. Two types of moving face stimuli were used: full-face versions (both spoken form and timing information available) and modified face versions (only timing information provided by peri-oral motion available). The results showed that the peri-oral timing information facilitated response time for speech and non-speech stimuli compared to a static face. An additional facilitatory effect was found for full-face versions compared to the timing condition this effect only occurred for speech stimuli. We propose the timing effect was due to cross-modal phase resetting the form effect to cross-modal priming.
Publisher: Elsevier BV
Date: 03-2012
DOI: 10.1016/J.CONCOG.2011.11.007
Abstract: The present study used masked repetition priming to examine whether face representations can be accessed without attention. Two experiments using a face recognition task (fame judgement) presented masked repetition and control primes in spatially unattended locations prior to target onset. Experiment 1 (n=20) used the same images as primes and as targets and Experiment 2 (n=17) used different images of the same in idual as primes and targets. Repetition priming was observed across both experiments regardless of whether spatial attention was cued to the location of the prime. Priming occurred for both famous and non-famous targets in Experiment 1 but was only reliable for famous targets in Experiment 2, suggesting that priming in Experiment 1 indexed access to view-specific representations whereas priming in Experiment 2 indexed access to view-invariant, abstract representations. Overall, the results indicate that subliminal access to abstract face representations does not rely on attention.
Publisher: Elsevier BV
Date: 09-2012
Publisher: Springer Nature Singapore
Date: 2023
Publisher: Informa UK Limited
Date: 04-1999
Publisher: Elsevier BV
Date: 11-2008
DOI: 10.1016/J.BRAINRES.2008.04.077
Abstract: We used whole-head magnetoencephalograpy (MEG) to record changes in neuromagnetic N100m responses generated in the left and right auditory cortex as a function of the match between visual and auditory speech signals. Stimuli were auditory-only (AO) and auditory-visual (AV) presentations of i/, /ti/ and /vi/. Three types of intensity matched auditory stimuli were used: intact speech (Normal), frequency band filtered speech (Band) and speech-shaped white noise (Noise). The behavioural task was to detect the /vi/ syllables which comprised 12% of stimuli. N100m responses were measured to averaged i/ and /ti/ stimuli. Behavioural data showed that identification of the stimuli was faster and more accurate for Normal than for Band stimuli, and for Band than for Noise stimuli. Reaction times were faster for AV than AO stimuli. MEG data showed that in the left hemisphere, N100m to both AO and AV stimuli was largest for the Normal, smaller for Band and smallest for Noise stimuli. In the right hemisphere, Normal and Band AO stimuli elicited N100m responses of quite similar litudes, but N100m litude to Noise was about half of that. There was a reduction in N100m for the AV compared to the AO conditions. The size of this reduction for each stimulus type was same in the left hemisphere but graded in the right (being largest to the Normal, smaller to the Band and smallest to the Noise stimuli). The N100m decrease for the Normal stimuli was significantly larger in the right than in the left hemisphere. We suggest that the effect of processing visual speech seen in the right hemisphere likely reflects suppression of the auditory response based on AV cues for place of articulation.
Publisher: Frontiers Media SA
Date: 2012
Publisher: Elsevier BV
Date: 06-2010
Publisher: SAGE Publications
Date: 2003
DOI: 10.1068/P3466
Abstract: We investigated audio-visual (AV) perceptual integration by examining the effect of seeing the speaker's synchronised moving face on masked-speech detection ability. Signal lification and higher-level cognitive accounts of an AV advantage were contrasted, the latter by varying whether participants knew the language of the speaker. An AV advantage was shown for sentences whose mid-to-high-frequency acoustic envelope was highly correlated with articulator movement, regardless of knowledge of the language. For low-correlation sentences, knowledge of the language had a large impact for participants with no knowledge of the language an AV inhibitory effect was found (providing support for reports of a compelling AV illusion). The results indicate a role for both sensory enhancement and higher-level cognitive factors in AV speech detection.
Publisher: Elsevier BV
Date: 07-2007
DOI: 10.1016/J.JECP.2007.01.006
Abstract: A masked priming procedure was used to explore developmental changes in the tuning of lexical word recognition processes. Lexical tuning was assessed by examining the degree of masked form priming and used two different types of prime-target lexical similarity: one letter different (e.g., rlay-->PLAY) and transposed letters (e.g., lpay-->PLAY). The performance of skilled adult readers was compared with that of developing readers in Grade 3. The same children were then tested again two years later, when they were in Grade 5. The skilled adult readers showed no form priming, indicating that their recognition mechanisms for these items had become finely tuned. In contrast, the Grade 3 readers showed substantial form priming effects for both measures of lexical similarity. When retested in Grade 5, the developing readers no longer showed significant one letter different priming, but transposed letter priming remained. In general, these results provide evidence for a transition from more broadly tuned to more finely tuned lexical recognition mechanisms and are interpreted in the context of models of word recognition.
Publisher: Elsevier BV
Date: 09-2013
DOI: 10.1016/J.BANDL.2013.06.008
Abstract: An important property of visual speech (movements of the lips and mouth) is that it generally begins before auditory speech. Research using brain-based paradigms has demonstrated that seeing visual speech speeds up the activation of the listener's auditory cortex but it is not clear whether these observed neural processes link to behaviour. It was hypothesized that the very early portion of visual speech (occurring before auditory speech) will allow listeners to predict the following auditory event and so facilitate the speed of speech perception. This was tested in the current behavioural experiments. Further, we tested whether the salience of the visual speech played a role in this speech facilitation effect (Experiment 1). We also determined the relative contributions that visual form (what) and temporal (when) cues made (Experiment 2). The results showed that visual speech cues facilitated response times and that this was based on form rather than temporal cues.
Publisher: Cambridge University Press (CUP)
Date: 28-06-2005
DOI: 10.1017/S0022226705003294
Abstract: Korean has a very complex phonology, with many interacting alternations. In a coronal-/i/ sequence, depending on the type of phonological boundary present, alternations such as palatalization, nasal insertion, nasal assimilation, coda neutralization, and intervocalic voicing can apply. This paper investigates how the phonological patterns of Korean affect processing of morphemes and words. Past research on languages such as English, German, Dutch, and Finnish has shown that listeners exploit syllable structure constraints in processing speech and segmenting it into words. The current study shows that in parsing speech, listeners also use much more complex patterns that relate the surface phonological string to various boundaries.
Publisher: ISCA
Date: 08-09-2016
Publisher: Elsevier BV
Date: 03-2014
Publisher: SAGE Publications
Date: 2010
DOI: 10.1068/P6483
Abstract: Parsing of information from the world into objects and events occurs in both the visual and auditory modalities. It has been suggested that visual and auditory scene perceptions involve similar principles of perceptual organisation. We investigated here cross-modal scene perception by determining whether an auditory stimulus could facilitate visual object segregation. Specifically, we examined whether the presentation of matched auditory speech would facilitate the detection of a point-light talking face amid point-light distractors. An adaptive staircase procedure (3-up–1-down rule) was used to estimate the 79% correct threshold in a two-alternative forced-choice procedure. To determine if different degrees of speech motion would show auditory influence of different sizes, two speech modes were tested (in quiet and Lombard speech). A facilitatory auditory effect on talking-face detection was found the size of this effect did not differ between the different speech modes.
Publisher: Acoustical Society of America (ASA)
Date: 04-2010
DOI: 10.1121/1.3353116
Abstract: The current study investigated the robustness of priming from a masked speech priming method introduced by Kouider and Dupoux [(2005). Psychol. Sci. 16, 617–625]. In this procedure, a compressed spoken prime is embedded in auditory masking stimuli and presented immediately prior to an uncompressed auditory target. The degree to which spoken stimuli could be compressed without significant data loss was first determined. Using this compression level, repetition and form priming were measured for the target words with a high versus low number of phonological neighbors. The results indicated that robust masked speech priming occurred only for word targets that had few neighbors.
Publisher: SAGE Publications
Date: 08-2004
DOI: 10.1080/02724980343000701
Abstract: The effects of viewing the face of the talker (visual speech) on the processing of clearly presented intact auditory stimuli were investigated using two measures likely to be sensitive to the articulatory motor actions produced in speaking. The aim of these experiments was to highlight the need for accounts of the effects of audio–visual (AV) speech that explicitly consider the properties of articulated action. The first experiment employed a syllable–monitoring task in which participants were required to monitor for target syllables within foreign carrier phrases. An AV effect was found in that seeing a talker's moving face (moving face condition) assisted in more accurate recognition (hits and correct rejections) of spoken syllables than of auditory–only still face (still face condition) presentations. The second experiment examined processing of spoken phrases by investigating whether an AV effect would be found for estimates of phrase duration. Two effects of seeing the moving face of the talker were found. First, the moving face condition had significantly longer duration estimates than the still face auditory–only condition. Second, estimates of auditory duration made in the moving face condition reliably correlated with the actual durations whereas those made in the still face auditory condition did not. The third experiment was carried out to determine whether the stronger correlation between estimated and actual duration in the moving face condition might have been due to generic properties of AV presentation. Experiment 3 employed the procedures of the second experiment but used stimuli that were not perceived as speech although they possessed the same timing cues as those of the speech stimuli of Experiment 2. It was found that simply presenting both auditory and visual timing information did not result in more reliable duration estimates. Further, when released from the speech context (used in Experiment 2), duration estimates for the auditory–only stimuli were significantly correlated with actual durations. In all, these results demonstrate that visual speech can assist in the analysis of clearly presented auditory stimuli in tasks concerned with information provided by viewing the production of an utterance. We suggest that these findings are consistent with there being a processing link between perception and action such that viewing a talker speaking will activate speech motor schemas in the perceiver.
Publisher: ISCA
Date: 08-09-2016
Publisher: Psychology Press
Date: 14-12-2010
Publisher: American Psychological Association (APA)
Date: 11-2008
DOI: 10.1037/A0013357
Abstract: The current research uses a novel methodology to examine the role of semantics in reading aloud. Participants were trained to read aloud 2 sets of novel words (i.e., nonwords such as bink): some with meanings (semantic) and some without (nonsemantic). A comparison of reading aloud performance between these 2 sets of novel words was used to provide an indicator of the importance of semantic information in reading aloud. In Experiment 1, in contrast to expectations, reading aloud performance was not better for novel words in the semantic condition. In Experiment 2, the training of novel words was modified to reflect more realistic steps of lexical acquisition: Reading aloud performance became faster and more accurate for novel words in the semantic condition, but only for novel words with inconsistent pronunciations. This semantic advantage for inconsistent novel words was again observed when a subset of participants from Experiment 2 was retested 6-12 months later (in Experiment 3). These findings provide support for a limited but significant role for semantics in the reading aloud process.
Publisher: Elsevier BV
Date: 07-2006
Publisher: Elsevier BV
Date: 09-2001
DOI: 10.1016/S0010-0277(01)00129-9
Abstract: Recently, Cestnick and Coltheart (Cognition 71 (1999) 231) have reported evidence of abnormal performance on the Ternus apparent motion task in dyslexics. We demonstrate that some aspects of their data may be accounted for by more frequent lapses of concentration in the dyslexic group than in controls. We then report on a study in which a modification of the Ternus procedure was employed to simplify the task and to control for the effects of inattention. The results suggest that dyslexics do genuinely differ from normal readers in their perceptual processing.
Publisher: Elsevier BV
Date: 12-2004
DOI: 10.1016/J.BANDL.2004.05.001
Abstract: The current research examined performance of good and poor readers of Thai on two tasks that assess sensitivity to dynamic visual displays. Readers of Thai, a complex alphabetic script that nonetheless has a regular orthography, were chosen in order to contrast patterns of performance with readers of Korean Hangul (a similarly regular language but one that has a simple visual format, see ). Thai poor readers were less sensitive than good readers on the two measures of dynamic visual processing they had higher thresholds for detecting coherent movement and required longer ISIs to report group movement in the Ternus task. These results differ from those for poor readers of Korean Hangul that found no relationship between visual processing thresholds and reading skill. This contrast suggests that the expression of visual processing problems in dyslexia is mediated by the format properties of the writing system and points to the need to consider such factors in formulating brain-behavior relationships.
Publisher: Acoustical Society of America (ASA)
Date: 12-2009
DOI: 10.1121/1.3250425
Abstract: This study investigated the degree to which two types of reduced auditory signals (cochlear implant simulations) and visual speech cues combined for speech identification. The auditory speech stimuli were filtered to have only litude envelope cues or both litude envelope and spectral cues and were presented with/without visual speech. In Experiment 1, IEEE sentences were presented in quiet and noise. For in-quiet presentation, speech identification was enhanced by the addition of both spectral and visual speech cues. Due to a ceiling effect, the degree to which these effects combined could not be determined. In noise, these facilitation effects were more marked and were additive. Experiment 2 examined consonant and vowel identification in the context of CVC or VCV syllables presented in noise. For consonants, both spectral and visual speech cues facilitated identification and these effects were additive. For vowels, the effect of combined cues was underadditive, with the effect of spectral cues reduced when presented with visual speech cues. Analysis indicated that without visual speech, spectral cues facilitated the transmission of place information and vowel height, whereas with visual speech, they facilitated lip rounding, with little impact on the transmission of place information.
Publisher: SAGE Publications
Date: 06-2015
Abstract: Three naming aloud experiments and a lexical decision (LD) experiment used masked priming to index the processing of written Thai vowels and tones. Thai allows for manipulation of the mapping between orthography and phonology not possible in other orthographies, for ex le, the use of consonants, vowels and tone markers in both horizontal and vertical orthographic positions (HOPs and VOPs). Experiment 1 showed that changing a vowel between prime and target slowed down target naming but changing a tone mark did not. Experiment 1 used an across item-design and a different number of HOPs in the way vowels and tones were specified. Experiment 2 used a within-item design and tested vowel and tone changes for both 2-HOP and 3-HOP targets separately. The 3-HOP words showed the same tone and vowel change effect as Experiment 1, whereas 2-HOP items did not. It was speculated that the 2-HOP result was due to the variable position of the vowel affecting priming. Experiment 3 employed a more stringent control over the 2-HOP vowel and tone items and found priming for the tone changes but not for vowel changes. The final experiment retested the items from Experiment 3 with the LD task and found no priming for the tone change items, indicating that the tone effect in Experiment 3 was due to processes involved in naming aloud. In all, the results supported the view that for naming a word, the development of tone information is slower than vowel information.
Publisher: Informa UK Limited
Date: 03-2007
DOI: 10.1080/02643290601025576
Abstract: The role of semantics in reading aloud remains controversial. To explore this issue, the current study examined the impact of semantic loss on reading-aloud performance in 7 patients with semantic dementia. The results revealed a heterogenous pattern of reading difficulties. Of the patients, 2 selectively made errors on inconsistent words (i.e., surface dyslexia), 4 had a generalized reading deficit with increased errors on consistent words, inconsistent words, and nonwords, while the remaining patient had relatively intact reading-aloud accuracy. All patients had longer reading latencies on real words than controls. The relationship between the reading and semantic deficits in the patients was examined at the item-specific level. This suggested that reading-aloud errors were related to the semantic impairment for inconsistent words but not consistent words. In contrast, semantic loss was related to longer latencies for both consistent and inconsistent words. These findings support models of reading that include a role for semantics in the reading-aloud process.
Publisher: Elsevier BV
Date: 05-2008
DOI: 10.1016/J.COGNITION.2007.07.016
Abstract: This study investigated whether masked priming is mediated by existing memory representations by determining whether nonwords targets would show repetition priming. To avoid the potential confound that nonword repetition priming would be obscured by a familiarity response bias, the standard lexical decision and naming tasks were modified to make targets unfamiliar. Participants were required to read a target string from right to left (i.e., "ECAF" should be read as "FACE") and then make a response. To examine if priming was based on lexical representations, repetition primes consisted of words when read forwards or backwards (e.g., "face", "ecaf") and nonwords (e.g., "pame", "emap"). Forward and backward primes were used to test if task instruction affected prime encoding. The lexical decision and naming tasks showed the same pattern of results: priming only occurred for forward primes with word targets (e.g., "face-ECAF"). Additional experiments to test if response priming affected the LDT indicated that the lexical status of the prime per se did not affect target responses. These results showed that the encoding of masked primes was unaffected by the novel task instruction and support the view that masked priming is due to the automatic triggering of pre-established computational processes based on stored information.
Publisher: Association for Research in Vision and Ophthalmology (ARVO)
Date: 15-02-2012
DOI: 10.1167/12.2.13
Abstract: The present study investigated the extent to which a face presented in the visual periphery is processed and whether such processing can be influenced by a recent encounter in central vision. To probe face processing, a series of studies was conducted in which participants classified the sex and identity of faces presented in central and peripheral vision. The results showed that when target faces had not been previously viewed in central vision, recognition in peripheral vision was limited whereas sex categorization was not. When faces were previously viewed in central vision, recognition in peripheral vision improved even with the pose, hairstyle, and lighting conditions of these faces changed. These results are discussed with regard to possible mechanisms unpinning this exposure effect.
Publisher: Acoustical Society of America (ASA)
Date: 02-2012
DOI: 10.1121/1.3676605
Abstract: This study investigated whether the production of prosodic focus and phrasing contrasts was modified when interlocutors could only hear each other [auditory only (AO)], compared to when they could hear and see each other [face to face (FTF)]. The prosodic characteristics of utterances produced by six talkers were examined using both acoustic and perceptual measures (ratings of the degree of focus or clarity of the statement-question contrast). The acoustic measures showed a range of differences between narrow focus and between phrasing contrasts and some of these differences were greater in the AO setting than the FTF one. The listener’s ratings of focus and phrasing showed a clear difference between the AO and FTF conditions, with perceptual attributes of both narrow focus and echoic question phrasing being rated as clearer in the AO condition. To explain these results it is proposed that talkers compensate for the lack of visual prosodic cues in the AO condition by taking extra care (relative to FTF conditions) to ensure the effective transmission of prosodic cues.
Publisher: Frontiers Media SA
Date: 31-08-2016
Publisher: SAGE Publications
Date: 12-2008
Abstract: To segment continuous speech into its component words, listeners make use of language rhythm because rhythm differs across languages, so do the segmentation procedures which listeners use. For each of stress-, syllable-and mora-based rhythmic structure, perceptual experiments have led to the discovery of corresponding segmentation procedures. In the case of mora-based rhythm, similar segmentation has been demonstrated in the otherwise unrelated languages Japanese and Telugu segmentation based on syllable rhythm, however, has been previously demonstrated only for European languages from the Romance family. We here report two target detection experiments in which Korean listeners, presented with speech in Korean and in French, displayed patterns of segmentation like those previously observed in analogous experiments with French listeners. The Korean listeners' accuracy in detecting word-initial target fragments in either language was significantly higher when the fragments corresponded exactly to a syllable in the input than when the fragments were smaller or larger than a syllable. We conclude that Korean and French listeners can call on similar procedures for segmenting speech, and we further propose that perceptual tests of speech segmentation provide a valuable accompaniment to acoustic analyses for establishing languages' rhythmic class membership.
Publisher: Elsevier BV
Date: 08-2004
Publisher: Elsevier BV
Date: 10-2004
Publisher: Elsevier BV
Date: 2002
DOI: 10.1016/S0028-3932(01)00092-6
Abstract: Medial temporal lobe (MTL) structures are implicated in forming conjunctions between events in order to form enduring relational memories these memories are not evident using direct measures with varieties of amnesic subjects. Extratemporal brain structures are thought to be responsible for preserved memories, which are sometimes detectable using indirect measures. The present study tests this theory of multiple memory systems by examining whether preserved learning can be demonstrated for relational material in MTL-disordered subjects using an indirect measure which minimises conscious mediation of performance. The subjects had undergone anterior temporal lobectomy for relief of temporal lobe epilepsy: left-sided (LATL) cases had a mild verbal amnesia and right-sided (RATL) cases had better verbal memory, forming a comparison group. A direct measure of verbal relational memory was provided by successive trials of cued recall in a specially-constructed paired associate learning task with arbitrarily paired words pairs consisted of either concrete or abstract words. LATL subjects performed worse than RATL subjects, and particularly so with abstract words. Following direct testing, memory for the pairings was measured indirectly using a masked recognition priming technique. RATL subjects showed savings in RT, demonstrating that masked priming can reveal evidence of the formation of conjunctions. Critically, LATL subjects showed no evidence of preserved learning with priming. Thus when MTL structures are damaged, relational memory appears to be affected without exception, consistent with the tenets of multiple memory systems theory.
Publisher: Public Library of Science (PLoS)
Date: 30-12-2022
DOI: 10.1371/JOURNAL.PONE.0279822
Abstract: The ability to recognise emotion from faces or voices appears to decline with advancing age. However, some studies have shown that emotion recognition of auditory-visual (AV) expressions is largely unaffected by age, i.e., older adults get a larger benefit from AV presentation than younger adults resulting in similar AV recognition levels. An issue with these studies is that they used well-recognised emotional expressions that are unlikely to generalise to real-life settings. To examine if an AV emotion recognition benefit generalizes across well and less well recognised stimuli, we conducted an emotion recognition study using expressions that had clear or unclear emotion information for both modalities, or clear visual, but unclear auditory information. Older (n = 30) and younger (n = 30) participants were tested on stimuli of anger, happiness, sadness, surprise, and disgust (expressed in spoken sentences) in auditory-only (AO), visual-only (VO), or AV format. Participants were required to respond by choosing one of 5 emotion options. Younger adults were more accurate in recognising emotions than older adults except for clear VO expressions. Younger adults showed an AV benefit even when unimodal recognition was poor. No such AV benefit was found for older adults indeed, AV was worse than VO recognition when AO recognition was poor. Analyses of confusion responses indicated that older adults generated more confusion responses that were common between AO and VO conditions, than younger adults. We propose that older adults’ poorer AV performance may be due to a combination of weak auditory emotion recognition and response uncertainty that resulted in a higher cognitive load.
Publisher: IEEE
Date: 09-2014
Publisher: Informa UK Limited
Date: 05-2009
Publisher: Elsevier BV
Date: 02-2014
Publisher: Psychology Press
Date: 02-06-2003
Publisher: SAGE Publications
Date: 10-2011
DOI: 10.1080/17470218.2011.573150
Abstract: Three experiments tested how the physical format and information content of forward and backward masks affected the extent of visual pattern masking. This involved using different types of forward and backward masks with target discrimination measured by percentage correct in the first experiment (with a fixed target duration) and by an adaptive threshold procedure in the last two. The rationale behind the manipulation of the content of the masks stemmed from masking theories emphasizing attentional and/or conceptual factors rather than visual ones. Experiment 1 used word masks and showed that masking was reduced (a masking reduction effect) when the forward and backward masks were the same word (although in different case) compared to when the masks were different words. Experiment 2 tested the extent to which a reduction in masking might occur due to the physical similarity between the forward and backward masks by comparing the effect of the same content of the masks in the same versus different case. The result showed a significant reduction in masking for same content masks but no significant effect of case. The last experiment examined whether the reduction in masking effect would be observed with nonword masks—that is, having no high-level representation. No reduction in masking was found from same compared to different nonword masks (Experiment 3). These results support the view that the conscious perception of a rapidly displayed target stimulus is in part determined by high-level perceptual/cognitive factors concerned with masking stimulus grouping and attention.
Publisher: Public Library of Science (PLoS)
Date: 21-10-2014
Publisher: Elsevier BV
Date: 05-2015
DOI: 10.1016/J.CONCOG.2015.02.016
Abstract: We examined if cross-modal priming (print to speech) was greater for participants who were aware of the presence of letters in the experiment. Experiment 1 determined that word primes displayed at 47ms were adequately masked. In Experiment 2 (a,b) with primes displayed at 47ms masked priming occurred for within-mode printed targets but not for spoken ones. Experiment 3, with spoken targets, presented primes at two different durations (59, 71ms) and priming was found for participants who reported seeing letters but not for those who did not. The results are discussed in terms how the link between prime and target representations might be strengthened even by cursory awareness of the prime and what this tells us about priming.
Publisher: Elsevier BV
Date: 08-2015
DOI: 10.1016/J.HUMOV.2015.05.009
Abstract: The present study examined the extent to which speech and manual gestures spontaneously entrain in a non-communicative task. Participants had to repeatedly utter nonsense /CV/ syllables while continuously moving the right index finger in flexion/extension. No instructions to coordinate were given. We manipulated the type of syllable uttered (/ba/ vs. /sa/), and vocalization (phonated vs. silent speech). Assuming principles of coordination dynamics, a stronger entrainment between the fingers oscillations and the jaw motion was predicted (1) for /ba/, due to expected larger litude of jaw motion and (2) in phonated speech, due to the auditory feedback. Fifteen out of twenty participants showed simple ratios of speech to finger cycles (1:1, 1:2 or 2:1). In contrast with our predictions, speech-gesture entrainment was stronger when vocalizing /sa/ than /ba/, also more widely distributed on an in-phase mode. Furthermore, results revealed a spatial anchoring and an increased temporal variability in jaw motion when producing /sa/. We suggested that this indicates a greater control of the speech articulators for /sa/, making the speech performance more receptive to environmental forces, resulting in the greater entrainment observed to gesture oscillations. The speech-gesture coordination was maintained in silent speech, suggesting a somatosensory basis for their endogenous coupling.
Publisher: American Speech Language Hearing Association
Date: 12-12-2022
DOI: 10.1044/2022_JSLHR-22-00218
Abstract: This study aimed to develop and test a measure of real-time continuous speech understanding to be used with natural dialogues. The measure was based on a category monitoring paradigm and employed five existing recordings of natural dialogues from which the different test categories and associated target words were derived. For each dialogue, a listener was first given a semantic category and asked to press a button as quickly as possible whenever they heard an instance of the category. We tested 63 younger adults, using five semantic categories (family, media, season, temperature, and travel) at three noise levels (in quiet, 0 dB, and −5 dB signal-to-noise ratio [SNR]). Performance was measured in terms of accuracy and response time. The results showed clear differences between the three noise conditions regardless of the semantic category. The peak of the response distribution was highest and earliest for the quiet condition and was reduced with decreasing SNR. The responses varied across categories, reflecting differences in the complexity of a given category or the typicality of the association between target words and their category. Broad categories and/or target words that were less directly associated with their category had decreased hit rates and increased response times. The results were discussed in terms of the sensitivity (hit rate) of the performance measure, as well as whether it picked up higher level semantic, context, and discourse properties of the dialogues. 0.23641/asha.21561681
Publisher: Elsevier BV
Date: 08-2019
DOI: 10.1016/J.CORTEX.2019.02.014
Abstract: Our study proposes a test of a key assumption of the most prominent model of consciousness - the global workspace (GWS) model (e.g., Baars, 2002, 2005, 2007 Dehaene & Naccache, 2001 Mudrik, Faivre, & Koch, 2014). This assumption is that multimodal integration requires consciousness however, few studies have explicitly tested if integration can occur between nonconscious information from different modalities. The proposed study examined whether a classic indicator of multimodal integration - the McGurk effect - can be elicited with subliminal auditory-visual speech stimuli. We used a masked speech priming paradigm developed by Kouider and Dupoux (2005) in conjunction with continuous flash suppression (CFS Tsuchiya & Koch, 2005), a binocular rivalry technique for presenting video stimuli subliminally. Applying these techniques together, we carried out two experiments in which participants categorised auditory syllable targets which were preceded by subliminal auditory-visual (AV) speech primes. Subliminal AV primes were either illusion-inducing (McGurk) or illusion-neutral (Incongruent) combinations of speech stimuli. In Experiment 1, the categorisation of the syllable target ("pa") was facilitated by the same syllable prime when it was part of a McGurk combination (auditory "pa" and visual "ka") but not when part of an Incongruent combination (auditory "pa" and visual "wa"). This dependency on specific AV combinations indicated a nonconscious AV interaction. Experiment 2 presented a different syllable target ("ta") which matched the predicted illusory outcome of the McGurk combination - here, both the McGurk combination (auditory "pa" and visual "ka") and the Incongruent combination (auditory "ta" and visual "ka") failed to facilitate target categorisation. The combined results of both Experiments demonstrate a type of nonconscious multimodal interaction that is distinct from integration - it allows unimodal information that is compatible for integration (i.e., McGurk combinations) to persist and influence later processes, but does not actually combine and alter that information. As the GWS model does not account for non-integrative multimodal interactions, this places some pressure on such models of consciousness.
Publisher: Elsevier BV
Date: 02-2008
Publisher: Elsevier BV
Date: 11-2003
Publisher: Informa UK Limited
Date: 11-06-2010
Publisher: Acoustical Society of America (ASA)
Date: 05-2014
DOI: 10.1121/1.4870701
Abstract: The primary aim of this study was to determine whether late French/English bilinguals are able to utilize knowledge of bilabial stop contrasts that exist in each of their separate languages to discriminate bilabial stop contrasts from a new language (Thai). Secondary aims were to determine associations between bilabial stop consonant production in the L1 and the L2, between language learning factors and production and discrimination, and to compare English bilinguals' and monolinguals' discrimination. Three Thai bilabial stop consonant pairs differentiated by Voice Onset Time (VOT) (combinations of [b], [p], and [ph]) were presented to 28 French-English bilinguals, 25 English-French bilinguals, and 43 English monolinguals in an AX discrimination task. It was hypothesized that L2 experience would facilitate discrimination of contrasts that were phonemic in the L2 but not in the L1 for bilinguals. Only limited support for this hypothesis was found. However, results indicate that high production proficiency bilinguals had higher discrimination of the phonemic L2 contrasts (non-phonemic in L1). Discrimination patterns indicate lasting L1 influence, with similarity between unknown foreign language contrasts and L1 contrasts influencing discrimination rates. Production results show evidence for L2 influence in the L1. Results are discussed in the context of current speech perception models.
Publisher: Elsevier BV
Date: 11-2015
Publisher: Elsevier BV
Date: 03-2012
DOI: 10.1016/J.COGNITION.2011.11.013
Abstract: Prosody can be expressed not only by modification to the timing, stress and intonation of auditory speech but also by modifying visual speech. Studies have shown that the production of visual cues to prosody is highly variable (both within and across speakers), however behavioural studies have shown that perceivers can effectively use such visual cues. The latter result suggests that people are sensitive to the type of prosody expressed despite cue variability. The current study investigated the extent to which perceivers can match visual cues to prosody from different speakers and from different face regions. Participants were presented two pairs of sentences (consisting of the same segmental content) and were required to decide which pair had the same prosody. Experiment 1 tested visual and auditory cues from the same speaker and Experiment 2 from different speakers. Experiment 3 used visual cues from the upper and the lower face of the same talker and Experiment 4 from different speakers. The results showed that perceivers could accurately match prosody even when signals were produced by different speakers. Furthermore, perceivers were able to match the prosodic cues both within and across modalities regardless of the face area presented. This ability to match prosody from very different visual cues suggests that perceivers cope with variation in the production of visual prosody by flexibly mapping specific tokens to abstract prosodic types.
Publisher: Elsevier BV
Date: 04-2011
DOI: 10.1016/J.JECP.2010.07.008
Abstract: The psycholinguistic status of lexical tones and phones is indexed via phonological and tonological awareness (PA and TA, respectively) using Thai speech. In Experiment 1 (Thai participants, alphabetic script and orthographically explicit phones/tones), PA was better than TA in children and primary school-educated adults, and TA improved to PA levels only in tertiary-educated adults. In Experiment 2 (Cantonese participants, logographic script and no orthographically explicit phones/tones), children and primary-educated adults had better PA than TA, and PA and TA were equivalent in tertiary-educated adults, but were nevertheless still below the level of their Thai counterparts. Experiment 3 (English-language participants, alphabetic script and nontonal) showed better PA than TA. Regression analyses showed that both TA and PA are predicted by reading ability for Thai children but by general nonorthographic age-related variables for Cantonese children, whereas for English children reading ability predicts PA but not TA. The results show a phone>tone perceptual advantage over both age and languages that is affected by availability of orthographically relevant information and metalinguistic maturity. More generally, both the perception and the psycholinguistic representation of phones and tones differ.
Publisher: Frontiers Media SA
Date: 2013
Publisher: SAGE Publications
Date: 2011
DOI: 10.1068/P6941
Abstract: Seeing the talker improves the intelligibility of speech degraded by noise (a visual speech benefit). Given that talkers exaggerate spoken articulation in noise, this set of two experiments examined whether the visual speech benefit was greater for speech produced in noise than in quiet. We first examined the extent to which spoken articulation was exaggerated in noise by measuring the motion of face markers as four people uttered 10 sentences either in quiet or in babble-speech noise (these renditions were also filmed). The tracking results showed that articulated motion in speech produced in noise was greater than that produced in quiet and was more highly correlated with speech acoustics. Speech intelligibility was tested in a second experiment using a speech-perception-in-noise task under auditory-visual and auditory-only conditions. The results showed that the visual speech benefit was greater for speech recorded in noise than for speech recorded in quiet. Furthermore, the amount of articulatory movement was related to performance on the perception task, indicating that the enhanced gestures made when speaking in noise function to make speech more intelligible.
Publisher: Elsevier BV
Date: 02-2016
DOI: 10.1016/J.CORTEX.2015.03.010
Abstract: We investigated whether internal models of the relationship between lip movements and corresponding speech sounds [Auditory-Visual (AV) speech] could be updated via experience. AV associations were indexed by early and late event related potentials (ERPs) and by oscillatory power and phase locking. Different AV experience was produced via a context manipulation. Participants were presented with valid (the conventional pairing) and invalid AV speech items in either a 'reliable' context (80% AVvalid items) or an 'unreliable' context (80% AVinvalid items). The results showed that for the reliable context, there was N1 facilitation for AV compared to auditory only speech. This N1 facilitation was not affected by AV validity. Later ERPs showed a difference in litude between valid and invalid AV speech and there was significant enhancement of power for valid versus invalid AV speech. These response patterns did not change over the context manipulation, suggesting that the internal models of AV speech were not updated by experience. The results also showed that the facilitation of N1 responses did not vary as a function of the salience of visual speech (as previously reported) in post-hoc analyses, it appeared instead that N1 facilitation varied according to the relative time of the acoustic onset, suggesting for AV events N1 may be more sensitive to the relationship of AV timing than form.
Publisher: Informa UK Limited
Date: 12-2002
No related organisations have been discovered for Chris Davis.
Start Date: 2002
End Date: 2004
Funder: Australian Research Council
View Funded ActivityStart Date: 2006
End Date: 2011
Funder: Australian Research Council
View Funded ActivityStart Date: 2003
End Date: 2005
Funder: Australian Research Council
View Funded ActivityStart Date: 2004
End Date: 2009
Funder: Australian Research Council
View Funded ActivityStart Date: 2003
End Date: 2003
Funder: Australian Research Council
View Funded ActivityStart Date: 01-2003
End Date: 08-2007
Amount: $215,000.00
Funder: Australian Research Council
View Funded ActivityStart Date: 09-2020
End Date: 09-2025
Amount: $411,603.00
Funder: Australian Research Council
View Funded ActivityStart Date: 01-2015
End Date: 12-2019
Amount: $305,700.00
Funder: Australian Research Council
View Funded ActivityStart Date: 03-2010
End Date: 08-2014
Amount: $200,000.00
Funder: Australian Research Council
View Funded ActivityStart Date: 09-2006
End Date: 10-2009
Amount: $130,000.00
Funder: Australian Research Council
View Funded ActivityStart Date: 03-2013
End Date: 03-2017
Amount: $438,000.00
Funder: Australian Research Council
View Funded ActivityStart Date: 10-2006
End Date: 07-2014
Amount: $3,400,000.00
Funder: Australian Research Council
View Funded ActivityStart Date: 09-2006
End Date: 01-2012
Amount: $639,000.00
Funder: Australian Research Council
View Funded ActivityStart Date: 12-2004
End Date: 12-2010
Amount: $2,000,000.00
Funder: Australian Research Council
View Funded ActivityStart Date: 01-2004
End Date: 11-2004
Amount: $20,000.00
Funder: Australian Research Council
View Funded Activity