Ändra sökning
Avgränsa sökresultatet
12 1 - 50 av 97
RefereraExporteraLänk till träfflistan
Permanent länk
Referera
Referensformat
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Annat format
Fler format
Språk
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Annat språk
Fler språk
Utmatningsformat
  • html
  • text
  • asciidoc
  • rtf
Träffar per sida
  • 5
  • 10
  • 20
  • 50
  • 100
  • 250
Sortering
  • Standard (Relevans)
  • Författare A-Ö
  • Författare Ö-A
  • Titel A-Ö
  • Titel Ö-A
  • Publikationstyp A-Ö
  • Publikationstyp Ö-A
  • Äldst först
  • Nyast först
  • Skapad (Äldst först)
  • Skapad (Nyast först)
  • Senast uppdaterad (Äldst först)
  • Senast uppdaterad (Nyast först)
  • Disputationsdatum (tidigaste först)
  • Disputationsdatum (senaste först)
  • Standard (Relevans)
  • Författare A-Ö
  • Författare Ö-A
  • Titel A-Ö
  • Titel Ö-A
  • Publikationstyp A-Ö
  • Publikationstyp Ö-A
  • Äldst först
  • Nyast först
  • Skapad (Äldst först)
  • Skapad (Nyast först)
  • Senast uppdaterad (Äldst först)
  • Senast uppdaterad (Nyast först)
  • Disputationsdatum (tidigaste först)
  • Disputationsdatum (senaste först)
Markera
Maxantalet träffar du kan exportera från sökgränssnittet är 250. Vid större uttag använd dig av utsökningar.
  • 1.
    Aare, Kätlin
    et al.
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik. University of Tartu, Estonia.
    Gilmartin, Emer
    Włodarczak, Marcin
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik.
    Lippus, Pärtel
    Heldner, Mattias
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik.
    Breath holds in chat and chunk phases of multiparty casual conversation2020Ingår i: Proceedings of Speech Prosody 2020, 2020, s. 779-783Konferensbidrag (Refereegranskat)
    Ladda ner fulltext (pdf)
    fulltext
  • 2.
    Aare, Kätlin
    et al.
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik, Avdelningen för fonetik.
    Lippus, Pärtel
    Włodarczak, Marcin
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik, Avdelningen för fonetik.
    Heldner, Mattias
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik, Avdelningen för fonetik.
    Creak in the respiratory cycle2018Ingår i: Proceedings of Interspeech 2018 / [ed] B. Yegnanarayana, The International Speech Communication Association (ISCA), 2018, s. 1408-1412Konferensbidrag (Refereegranskat)
    Abstract [en]

    Creakiness is a well-known turn-taking cue and has been observed to systematically accompany phrase and turn ends in several languages. In Estonian, creaky voice is frequently used by all speakers without any obvious evidence for its systematic use as a turn-taking cue. Rather, it signals a lack of prominence and is favored by lengthening and later timing in phrases. In this paper, we analyze the occurrence of creak with respect to properties of the respiratory cycle. We show that creak is more likely to accompany longer exhalations. Furthermore, the results suggest there is little difference in lung volume values regardless of the presence of creak, indicating that creaky voice might be employed to preserve air over the course of longer utterances. We discuss the results in connection to processes of speech planning in spontaneous speech.

    Ladda ner fulltext (pdf)
    fulltext
  • 3.
    Aare, Kätlin
    et al.
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik, Avdelningen för fonetik.
    Włodarczak, Marcin
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik, Avdelningen för fonetik.
    Heldner, Mattias
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik, Avdelningen för fonetik.
    Backchannels and breathing2014Ingår i: Proceedings from FONETIK 2014: Stockholm, June 9-11, 2014 / [ed] Mattias Heldner, Stockholm: Department of Linguistics, Stockholm University , 2014, s. 47-52Konferensbidrag (Övrigt vetenskapligt)
    Abstract [en]

    The present study investigated the timing of backchannel onsets within speaker’s own and dialogue partner’s breathing cycle in two spontaneous conversations in Estonian. Results indicate that backchannels are mainly produced near the beginning, but also in the second half of the speaker’s exhalation phase. A similar tendency was observed in short non-backchannel utterances, indicating that timing of backchannels might be determined by their duration rather than their pragmatic function. By contrast, longer non-backchannel utterances were initiated almost exclusively right at the beginning of the exhalation. As expected, backchannels in the conversation partner’s breathing cycle occurred predominantly towards the end of the exhalation or at the beginning of the inhalation. 

    Ladda ner fulltext (pdf)
    Backchannels and breathing
  • 4.
    Aare, Kätlin
    et al.
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik. University of Tartu, Estonia.
    Włodarczak, Marcin
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik.
    Heldner, Mattias
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik.
    Breath holds in spontaneous speech2019Ingår i: Eesti ja soome-ugri keeleteaduse ajakiri, ISSN 1736-8987, E-ISSN 2228-1339, Vol. 10, nr 1, s. 13-34Artikel i tidskrift (Refereegranskat)
    Abstract [en]

    This article provides a first quantitative overview of the timing and volume-related properties of breath holds in spontaneous conversations. Firstly, we investigate breath holds based on their position within the coinciding respiratory interval amplitude. Secondly, we investigate breath holds based on their timing within the respiratory intervals and in relation to communicative activity following breath holds. We hypothesise that breath holds occur in different regions of the lung capacity range and at different times during the respiratory phase, depending on the conversational and physiological activity following breath holds. The results suggest there is not only considerable variation in both the time and lung capacity scales, but detectable differences are also present in breath holding characteristics involving laughter and speech preparation, while breath holds coinciding with swallowing are difficult to separate from the rest of the data based on temporal and volume information alone.

  • 5.
    Aare, Kätlin
    et al.
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik, Avdelningen för fonetik.
    Włodarczak, Marcin
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik, Avdelningen för fonetik.
    Heldner, Mattias
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik, Avdelningen för fonetik.
    Inhalation amplitude and turn-taking in spontaneous Estonian conversations2015Ingår i: Proceedings from Fonetik 2015 Lund, June 8-10, 2015 / [ed] Malin Svensson Lundmark, Gilbert Ambrazaitis, Joost van de Weijer, Lund: Lund University , 2015, s. 1-5Konferensbidrag (Övrigt vetenskapligt)
    Abstract [en]

    This study explores the relationship between inhalation amplitude and turn management in four approximately 20 minute long spontaneous multiparty conversations in Estonian. The main focus of interest is whether inhalation amplitude is greater before turn onset than in the following inhalations within the same speaking turn. The results show that inhalations directly before turn onset are greater in amplitude than those later in the turn. The difference seems to be realized by ending the inhalation at a greater lung volume value, whereas the initial lung volume before inhalation onset remains roughly the same across a single turn. The findings suggest that the increased inhalation amplitude could function as a cue for claiming the conversational floor.

    Ladda ner fulltext (pdf)
    fulltext
  • 6.
    Bell, Linda
    et al.
    TeliaSonera (R and D).
    Boye, Johan
    TeliaSonera (R and D).
    Gustafson, Joakim
    TeliaSonera (R&D).
    Heldner, Mattias
    TeliaSonera (R&D).
    Lindström, Anders
    TeliaSonera (R and D).
    Wirén, Mats
    TeliaSonera (R&D).
    The Swedish NICE Corpus – Spoken dialogues between children and embodied characters in a computer game scenario2005Ingår i: Proceedings Interspeech 2005 - Eurospeech: 9th European Conference on Speech Communication and Technology, Lisbon, Portugal: ISCA , 2005, s. 2765-2768Konferensbidrag (Refereegranskat)
    Abstract [en]

    This article describes the collection and analysis of a Swedish database of spontaneous and unconstrained children-machine dialogues. The Swedish NICE corpus consists of spoken dialogues between children aged 8 to 15 and embodied fairytale characters in a computer game scenario. Compared to previously collected corpora of children's computer-directed speech, the Swedish NICE corpus contains extended interactions, including three-party conversation, in which the young users used spoken dialogue as the primary means of progression in the game.

    Ladda ner fulltext (pdf)
    The Swedish NICE Corpus – Spoken dialogues between children and embodied characters in a computer game scenario
  • 7.
    Bell, Linda
    et al.
    KTH Speech, Music and Hearing.
    Gustafson, Joakim
    KTH Speech, Music and Hearing.
    Heldner, Mattias
    KTH Speech, Music and Hearing.
    Prosodic adaptation in human-computer interaction2003Ingår i: Proceedings ICPhS 2003, Barcelona, Spain: ISCA , 2003, s. 2453-2456Konferensbidrag (Refereegranskat)
    Abstract [en]

    State-of-the-art speech recognizers are trained on predominantly normal speech and have difficulties handling either exceedingly slow and hyperarticulated or fast and sloppy speech. Explicitly instructing users on how to speak, however, can make the human–computer interaction stilted and unnatural. If it is possible to affect users’ speaking rate while maintaining the naturalness of the dialogue, this could prove useful in the development of future human–computer interfaces. Users could thus be subtly influenced to adapt their speech to better match the current capabilities of the system, so that errors can be reduced and the overall quality of the human–computer interaction is improved. At the same time, speakers are allowed to express themselves freely and naturally. In this article, we investigate whether people adapt their speech as they interact with an animated character in a simulated spoken dialogue system. A user experiment involving 16 subjects was performed to examine whether people who speak with a simulated dialogue system adapt their speaking rate to that of the system. The experiment confirmed that the users adapted to the speaking rate of the system, and no subjects afterwards seemed to be aware they had been affected in this way. Another finding was that speakers varied their speaking rate substantially in the course of the dialogue. In particular, problematic sequences where subjects had to repeat or rephrase the same utterance several times elicited slower speech.

  • 8.
    Berger, Alexandra
    et al.
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik, Avdelningen för fonetik. Karolinska Institutet, Sweden.
    Hedström Lindenhäll, Rosanna
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik, Avdelningen för fonetik. Karolinska Institutet, Sweden.
    Heldner, Mattias
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik, Avdelningen för fonetik.
    Karlsson, Sofia
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik, Avdelningen för fonetik. Karolinska Institutet, Sweden.
    Nyberg Pergament, Sarah
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik, Avdelningen för fonetik. Karolinska Institutet, Sweden.
    Vojnovic, Ivan
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik, Avdelningen för fonetik. Karolinska Institutet, Sweden.
    Voices after midnight: How a night out affects voice quality2014Ingår i: Proceedings from FONETIK 2014: Stockholm, June 9-11, 2014 / [ed] Mattias Heldner, Stockholm: Department of Linguistics, Stockholm University , 2014, s. 1-4Konferensbidrag (Övrigt vetenskapligt)
    Abstract [en]

    This study aimed to investigate how different parameters of the voice (jitter, shimmer, LTAS and mean pitch) are affected by a late night out. Three recordings were made: one early evening before the night out, one after midnight, and one on the next day. Each recording consisted of a one minute reading and prolonged vowels. Five students took part in the experiment. Results varied among the participants, but some patterns were noticeable in all parameters. A trend towards increased mean pitch during the second recording was observed among four of the subjects. Somewhat unexpectedly, jitter and shimmer decreased between the first and second recordings and increased in the third one. Due to the lack of ethical testing, only a small number of participants were included. A larger sample is suggested for future research in order to generalize results.

    Ladda ner fulltext (pdf)
    Voices after midnight: How a night out affects voice quality
  • 9.
    Beskow, Jonas
    et al.
    KTH Speech, Music and Hearing.
    Carlson, Rolf
    KTH Speech, Music and Hearing.
    Edlund, Jens
    KTH Speech, Music and Hearing.
    Granström, Björn
    KTH Speech, Music and Hearing.
    Heldner, Mattias
    KTH Speech, Music and Hearing.
    Hjalmarsson, Anna
    KTH Speech, Music and Hearing.
    Skantze, Gabriel
    KTH Speech, Music and Hearing.
    Multimodal Interaction Control2009Ingår i: Computers in the Human Interaction Loop / [ed] Waibel, Alex and Stiefelhagen, Rainer, Berlin/Heidelberg: Springer Berlin/Heidelberg, 2009, s. 143-158Kapitel i bok, del av antologi (Refereegranskat)
  • 10.
    Beskow, Jonas
    et al.
    KTH Speech, Music and Hearing.
    Edlund, Jens
    KTH Speech, Music and Hearing.
    Gustafson, Joakim
    KTH Speech, Music and Hearing.
    Heldner, Mattias
    KTH Speech, Music and Hearing.
    Hjalmarsson, Anna
    KTH Speech, Music and Hearing.
    House, David
    KTH Speech, Music and Hearing.
    Research focus: Interactional aspects of spoken face-to-face communication2010Ingår i: Proceedings from Fonetik 2010, Lund: Lund University , 2010, s. 7-10Konferensbidrag (Övrigt vetenskapligt)
  • 11.
    Edlund, Jens
    et al.
    KTH Speech, Music and Hearing.
    Alexandersson, Simon
    Beskow, Jonas
    KTH Speech, Music and Hearing.
    Gustavsson, Lisa
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik, Avdelningen för fonetik.
    Heldner, Mattias
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik, Avdelningen för fonetik.
    Hjalmarsson, Anna
    KTH Speech, Music and Hearing.
    Kallioinen, Petter
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik, Avdelningen för fonetik.
    Marklund, Ellen
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik, Avdelningen för fonetik.
    3rd party observer gaze as a continuous measure of dialogue flow2012Konferensbidrag (Refereegranskat)
    Abstract [en]

    We present an attempt at using 3rd party observer gaze to get a measure of how appropriate each segment in a dialogue is for a speaker change. The method is a step away from the current dependency of speaker turns or talkspurts towards a more general view of speaker changes. We show that 3rd party observers do indeed largely look at the same thing (the speaker), and how this can be captured and utilized to provide insights into human communication. In addition, the results also suggest that there might be differences in the distribution of 3rd party observer gaze depending on how information-rich an utterance is. 

    Ladda ner fulltext (pdf)
    fulltext
  • 12.
    Edlund, Jens
    et al.
    KTH Speech, Music and Hearing.
    Gustafson, Joakim
    KTH Speech, Music and Hearing.
    Heldner, Mattias
    KTH Speech, Music and Hearing.
    Hjalmarsson, Anna
    KTH Speech, Music and Hearing.
    Towards human-like spoken dialogue systems2008Ingår i: Speech Communication, ISSN 0167-6393, E-ISSN 1872-7182, Vol. 50, nr 8-9, s. 630-645Artikel i tidskrift (Refereegranskat)
    Abstract [en]

    This paper presents an overview of methods that can be used to collect and analyse data on user responses to spoken dialogue system components intended to increase human-likeness, and to evaluate how well the components succeed in reaching that goal. Wizard-of-Oz variations, human-human data manipulation, and micro-domains are discussed ill this context, as is the use of third-party reviewers to get a measure of the degree of human-likeness. We also present the two-way mimicry target, a model for measuring how well a human-computer dialogue mimics or replicates some aspect of human-human dialogue, including human flaws and inconsistencies. Although we have added a measure of innovation, none of the techniques is new in its entirely. Taken together and described from a human-likeness perspective, however, they form a set of tools that may widen the path towards human-like spoken dialogue systems.

  • 13.
    Edlund, Jens
    et al.
    KTH Speech, Music and Hearing.
    Heldner, Mattias
    KTH Speech, Music and Hearing.
    Exploring prosody in interaction control2005Ingår i: Phonetica, ISSN 0031-8388, E-ISSN 1423-0321, Vol. 62, nr 2-4, s. 215-226Artikel i tidskrift (Refereegranskat)
    Abstract [en]

    This paper investigates prosodic aspects of turn-taking in conversation with a view to improving the efficiency of identifying relevant places at which a machine can legitimately begin to talk to a human interlocutor. It examines the relationship between interaction control, the communicative function of which is to regulate the flow of information between interlocutors, and its phonetic manifestation. Specifically, the listener's perception of such interaction control phenomena is modelled. Algorithms for automatic online extraction of prosodic phenomena liable to be relevant for interaction control, such as silent pauses and intonation patterns, are presented and evaluated in experiments using Swedish map task data. We show that the automatically extracted prosodic features can be used to avoid many of the places where current dialogue systems run the risk of interrupting their users, as well as to identify suitable places to take the turn.

  • 14.
    Edlund, Jens
    et al.
    KTH Speech, Music and Hearing.
    Heldner, Mattias
    KTH Speech, Music and Hearing.
    /nailon/ – software for online analysis of prosody2006Ingår i: Proceedings Interspeech 2006, Pittsburgh, PA, USA: ISCA , 2006, s. 2022-2025Konferensbidrag (Refereegranskat)
  • 15.
    Edlund, Jens
    et al.
    KTH Speech, Music and Hearing.
    Heldner, Mattias
    KTH Speech, Music and Hearing.
    Underpinning /nailon/: automatic estimation of pitch range and speaker relative pitch2007Ingår i: Speaker Classification II / [ed] Müller, Christian, Berlin/Heidelberg, Germany: Springer Berlin/Heidelberg, 2007, s. 229-242Kapitel i bok, del av antologi (Refereegranskat)
  • 16.
    Edlund, Jens
    et al.
    KTH Speech, Music and Hearing.
    Heldner, Mattias
    KTH Speech, Music and Hearing.
    Al Moubayed, Samer
    KTH Speech, Music and Hearing.
    Gravano, Agustìn
    Hirschberg, Julia
    Columbia University Computer Science.
    Very short utterances in conversation2010Ingår i: Proceedings from Fonetik 2010, Lund: Lund University , 2010, s. 11-16Konferensbidrag (Övrigt vetenskapligt)
  • 17.
    Edlund, Jens
    et al.
    KTH Speech, Music and Hearing.
    Heldner, Mattias
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik, Avdelningen för fonetik.
    Gustafson, Joakim
    KTH Speech, Music and Hearing.
    On the effect of the acoustic environment on the accuracy of perception of speaker orientation from auditory cues alone2012Ingår i: INTERSPEECH 2012: vol.2, Portland, USA: Curran Associates, Inc. , 2012, s. 1482-1485Konferensbidrag (Refereegranskat)
    Abstract [en]

    The ability of people, and of machines, to determine the position of a sound source in a room is well studied. The related ability to determine the orientation of a directed sound source, on the other hand, is not, but the few studies there are show people to be surprisingly skilled at it. This has bearing for studies of face-to- face interaction and of embodied spoken dialogue systems, as sound source orientation of a speaker is connected to the head pose of the speaker, which is meaningful in a number of ways. The feature most often implicated for detection of sound source orientation is the inter-aural level difference - a feature which it is assumed is more easily exploited in anechoic chambers than in everyday surroundings. We expand here on our previous studies and compare detection of speaker orientation within and outside of the anechoic chamber. Our results show that listeners find the task easier, rather than harder, in everyday surroundings, which suggests that inter-aural level differences is not the only feature at play. 

    Ladda ner fulltext (pdf)
    fulltext
  • 18.
    Edlund, Jens
    et al.
    KTH Speech, Music and Hearing.
    Heldner, Mattias
    KTH Speech, Music and Hearing.
    Gustafson, Joakim
    KTH Speech, Music and Hearing.
    Utterance segmentation and turn-taking in spoken dialogue systems2005Ingår i: Sprachtechnologie, mobile kommunikation und linguistische ressourcen, Frankfurt am Main: Peter Lang Publishing Group, 2005, s. 576-587Kapitel i bok, del av antologi (Refereegranskat)
  • 19.
    Edlund, Jens
    et al.
    KTH Speech, Music and Hearing.
    Heldner, Mattias
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik, Avdelningen för fonetik.
    Gustafson, Joakim
    KTH Speech, Music and Hearing.
    Who am I speaking at? Perceiving the head orientation of speakers from acoustic cues alone2012Ingår i: LREC Workshop on Multimodal Corpora for Machine Learning, Istanbul, Turkey: LREC , 2012Kapitel i bok, del av antologi (Refereegranskat)
    Abstract [en]

    The ability of people, and of machines, to determine the position of a sound source in a room is well studied. The related ability to determine the orientation of a directed sound source, on the other hand, is not, but the few studies there are show people to be surprisingly skilled at it. This has bearing for studies of face-to-face interaction and of embodied spoken dialogue systems, as sound source orientation of a speaker is connected to the head pose of the speaker, which is meaningful in a number of ways. We describe in passing some preliminary findings that led us onto this line of investigation, and in detail a study in which we extend an experiment design intended to measure perception of gaze direction to test instead for perception of sound source orientation. The results corroborate those of previous studies, and further show that people are very good at performing this skill outside of studio conditions as well. 

    Ladda ner fulltext (pdf)
    fulltext
  • 20.
    Edlund, Jens
    et al.
    KTH Speech, Music and Hearing.
    Heldner, Mattias
    KTH Speech, Music and Hearing.
    Hirschberg, Julia
    Columbia University Computer Science.
    Pause and gap length in face-to-face interaction2009Ingår i: Proceedings of Interspeech 2009, Brighton, UK: ISCA , 2009, s. 2779-2782Konferensbidrag (Refereegranskat)
  • 21.
    Edlund, Jens
    et al.
    KTH Speech, Music and Hearing.
    Heldner, Mattias
    KTH Speech, Music and Hearing.
    Pelcé, Antoine
    KTH Speech, Music and Hearing.
    Prosodic features of very short utterances in dialogue2009Ingår i: Nordic Prosody: Proceedings of the Xth Conference, Helsinki 2008, Frankfurt am Main: Peter Lang Publishing Group, 2009, s. 57-68Konferensbidrag (Refereegranskat)
  • 22.
    Edlund, Jens
    et al.
    KTH Speech, Music and Hearing.
    Heldner, Mattias
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik, Avdelningen för fonetik.
    Włodarczak, Marcin
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik, Avdelningen för fonetik.
    Catching wind of multiparty conversation2014Ingår i: Proceedings of Multimodal Corpora: Combining applied and basic research targets (MMC 2014) / [ed] Jens Edlund, Dirk Heylen, Patrizia Paggio, Reykjavik, Iceland: European Language Resources Association , 2014, s. 35-36Kapitel i bok, del av antologi (Övrigt vetenskapligt)
    Abstract [en]

    The paper describes the design of a novel multimodal corpus of spontaneous multiparty conversations in Swedish. The corpus is collected with the primary goal of investigating the role of breathing and its perceptual cues for interactive control of interaction. Physiological correlates of breathing are captured by means of respiratory belts, which measure changes in cross sectional area of the rib cage and the abdomen. Additionally, auditory and visual correlates of breathing are recorded in parallel to the actual conversations. The corpus allows studying respiratory mechanisms underlying organisation of spontaneous conversation, especially in connection with turn management. As such, it is a valuable resource both for fundamental research and speech techonology applications.

    Ladda ner fulltext (pdf)
    fulltext
  • 23.
    Edlund, Jens
    et al.
    KTH Speech, Music and Hearing.
    Heldner, Mattias
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik. Department of Linguistics.
    Włodarczak, Marcin
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik.
    Is breathing prosody?2014Ingår i: International Symposium on Prosody to Commemorate Gösta Bruce, Lund: Lund University , 2014Konferensbidrag (Övrigt vetenskapligt)
    Abstract [en]

    Even though we may not be aware of it, much breathing in face-to-face conversation is both clearly audible and visible. Consequently, it has been suggested that respiratory activity is used in the joint coordination of conversational flow. For instance, it has been claimed that inhalation is an interactionally salient cue to speech initiation, that exhalation is a turn yielding device, and that breath holding is a marker of turn incompleteness (e.g. Local & Kelly, 1986; Schegloff, 1996). So far, however, few studies have addressed the interactional aspects of breathing (one notable exeption is McFarland, 2001). In this poster, we will describe our ongoing efforts to fill this gap. We will present the design of a novel corpus of respiratory activity in spontaneous multiparty face-to-face conversations in Swedish. The corpus will contain physiological measurements relevant to breathing, high-quality audio, and video. Minimally, the corpus will be annotated with interactional events derived from voice activity detection and (semi-) automatically detected inhalation and exhalation events in the respiratory data. We will also present initial analyses of the material collected. The question is whether breathing is prosody and relevant to this symposium? What we do know is that the turntaking phenomena that of particular interest to us are closely (almost by definition) related to several prosodic phenomena, and in particular to those associated with prosodic phrasing, grouping and boundaries. Thus, we will learn more about respiratory activity in phrasing (and the like) through analyses of breathing in conversation. References Local, John K., & Kelly, John. (1986). Projection and 'silences': Notes on phonetic and conversational structure. Human Studies, 9, 185-204. McFarland, David H. (2001). Respiratory markers of conversational interaction. Journal of Speech, Language, and Hearing Research, 44, 128-143. Schegloff, E. A. (1996). Turn organization: One intersection of grammar and interaction. In E. Ochs, E. A. Schegloff & S. A. Thompson (Eds.), Interaction and Grammar (pp. 52-133), Cambridge: Cambridge University Press.

  • 24.
    Eriksson, Anders
    et al.
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik, Avdelningen för fonetik.
    Bertinetto, Pier Marco
    Heldner, Mattias
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik, Avdelningen för fonetik.
    Nodari, Rosalba
    Lenoci, Giovanna
    The Acoustics of Lexical Stress in Italian as a Function of Stress Level and Speaking Style2016Ingår i: Proceedings of Interspeech 2016 / [ed] Nelson Morgan, The International Speech Communication Association (ISCA), 2016, s. 1059-1063Konferensbidrag (Refereegranskat)
    Abstract [en]

    The study is part of a series of studies, describing the acoustics of lexical stress in a way that should be applicable to any language. The present database of recordings includes Brazilian Portuguese, English, Estonian, German, French, Italian and Swedish. The acoustic parameters examined are F0-level, F0- variation, Duration, and Spectral Emphasis. Values for these parameters, computed for all vowels (a little over 24000 vowels for Italian), are the data upon which the analyses are based. All parameters are examined with respect to their correlation with Stress (primary, secondary, unstressed) and speaking Style (wordlist reading, phrase reading, spontaneous speech) and Sex of the speaker (female, male). For Italian Duration was found to be the dominant factor by a wide margin, in agreement with previous studies. Spectral Emphasis was the second most important factor. Spectral Emphasis has not been studied previously for Italian but intensity, a related parameter, has been shown to correlate with stress. F0-level was also significantly correlated but not to the same degree. Speaker Sex turned out as significant in many comparisons. The differences were, however, mainly a function of the degree to which a given parameter was used, not how it was used to signal lexical stress contrasts. 

    Ladda ner fulltext (pdf)
    fulltext
  • 25.
    Eriksson, Anders
    et al.
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik, Avdelningen för fonetik.
    Heldner, Mattias
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik, Avdelningen för fonetik.
    The acoustics of word stress in English as a function of stress level and speaking style2015Ingår i: 16th Annual Conference of the International Speech Communication Association (INTERSPEECH 2015): Speech Beyond Speech Towards a Better Understanding of the Most Important Biosignal, 2015, s. 41-45Konferensbidrag (Refereegranskat)
    Abstract [en]

    This study of lexical stress in English is part of a series of studies, the goal of which is to describe the acoustics of lexical stress for a number of typologically different languages. When fully developed the methodology should be applicable to any language. The database of recordings so far includes Brazilian Portuguese, English (U.K.), Estonian, German, French, Italian and Swedish. The acoustic parameters examined are f0-level, f0-variation, Duration, and Spectral Emphasis. Values for these parameters, computed for all vowels, are the data upon which the analyses are based. All parameters are tested with respect to their correlation with stress level (primary, secondary, unstressed) and speaking style (wordlist reading, phrase reading, spontaneous speech). For the English data, the most robust results concerning stress level are found for Duration and Spectral Emphasis. f0-level is also significantly correlated but not quite to the same degree. The acoustic effect of phonological secondary stress was significantly different from primary stress only for Duration. In the statistical tests, speaker sex turned out as significant in most cases. Detailed examination showed, however, that the difference was mainly in the degree to which a given parameter was used, not how it was used to signal lexical stress contrasts. 

    Ladda ner fulltext (pdf)
    fulltext
  • 26.
    Gustafson, Joakim
    et al.
    KTH Speech, Music and Hearing.
    Heldner, Mattias
    KTH Speech, Music and Hearing.
    Edlund, Jens
    KTH Speech, Music and Hearing.
    Potential benefits of human-like dialogue behaviour in the call routing domain2008Ingår i: Perception in Multimodal Dialogue Systems, Berlin/Heidelberg, Germany: Springer Berlin/Heidelberg, 2008, s. 240-251Kapitel i bok, del av antologi (Refereegranskat)
    Abstract [en]

    This paper presents a Wizard-of-Oz (Woz) experiment in the call routing domain that took place during the development of a call routing system for the TeliaSonera residential customer care in Sweden. A corpus of 42,000 calls was used as a basis for identifying problematic dialogues and the strategies used by operators to overcome the problems. A new Woz recording was made, implementing some of these strategies. The collected data is described and discussed with a view to explore the possible benefits of more human-like dialogue behaviour in call routing applications.

  • 27. Hammarsten, Jonna
    et al.
    Harris, Roxanne
    Henriksson, Nilla
    Pano, Isabelle
    Heldner, Mattias
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik, Avdelningen för fonetik.
    Włodarczak, Marcin
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik, Avdelningen för fonetik.
    Temporal aspects of breathing and turn-taking in Swedish multiparty conversations2015Ingår i: Proceedings from Fonetik 2015 / [ed] Malin Svensson Lundmark, Gilbert Ambrazaitis, Joost van de Weijer, Lund: Centre for Languages and Literature, 2015, s. 47-50Konferensbidrag (Övrigt vetenskapligt)
    Abstract [en]

    Interlocutors use various signals to make conversations flow smoothly. Recent research has shown that respiration is one of the signals used to indicate the intention to start speaking. In this study, we investigate whether inhalation duration and speech onset delay within one’s own turn differ from when a new turn is initiated. Respiratory activity was recorded in two three-party conversations using Respiratory Inductance Plethysmography. Inhalations were categorised depending on whether they coincided with within-speaker silences or with between- speaker silences. Results showed that within-turn inhalation durations were shorter than inhalations preceding new turns. Similarly, speech onset delays were shorter within turns than before new turns. Both these results suggest that speakers ‘speed up’ preparation for speech inside turns, probably to indicate that they intend to continue. 

    Ladda ner fulltext (pdf)
    fulltext
  • 28.
    Heldner, Mattias
    KTH Speech, Music and Hearing.
    Detection thresholds for gaps, overlaps and no-gap-no-overlaps2011Ingår i: Journal of the Acoustical Society of America, ISSN 0001-4966, Vol. 130, nr 1, s. 508-513Artikel i tidskrift (Refereegranskat)
    Abstract [en]

    Detection thresholds for gaps and overlaps, that is acoustic and perceived silences and stretches of overlapping speech in speaker changes, were determined. Subliminal gaps and overlaps were cate- gorized as no-gap-no-overlaps. The established gap and overlap detection thresholds both corre- sponded to the duration of a long vowel, or about 120 ms. These detection thresholds are valuable for mapping the perceptual speaker change categories gaps, overlaps, and no-gap-no-overlaps into the acoustic domain. Furthermore, the detection thresholds allow generation and understanding of gaps, overlaps, and no-gap-no-overlaps in human-like spoken dialogue systems.

  • 29.
    Heldner, Mattias
    Umeå University, Department of Philosophy and Linguistics.
    Is an F0-rise a necessary or a sufficient cue to perceived focus in Swedish?1998Ingår i: Nordic prosody: prodeedings of the VIIth conference, Joensuu 1996 / [ed] Stefan Werner, Frankfurt am Main: Peter Lang Publishing Group, 1998, s. 109-125Konferensbidrag (Refereegranskat)
  • 30.
    Heldner, Mattias
    Umeå University, Department of Philosophy and Linguistics.
    On the non-linear lengthening of focally accented Swedish words2001Ingår i: Nordic Prosody: proceedings of the VIIIth Conference, Trondheim 2000 / [ed] Wim A. van Dommelen, Thorstein Fretheim, Frankfurt am Main: Peter Lang Publishing Group, 2001, s. 103-112Konferensbidrag (Refereegranskat)
  • 31.
    Heldner, Mattias
    KTH Speech, Music and Hearing.
    On the reliability of overall intensity and spectral emphasis as acoustic correlates of focal accents in Swedish2003Ingår i: Journal of Phonetics, ISSN 0095-4470, E-ISSN 1095-8576, Vol. 31, nr 1, s. 39-62Artikel i tidskrift (Refereegranskat)
    Abstract [en]

    This study shows that increases in overall intensity and spectral emphasis are reliable acoustic correlates of focal accents in Swedish. They are both reliable in the sense that there are statistically significant differences between focally accented words and nonfocal ones for a variety of words, in any position of the phrase and for all speakers in the analyzed materials, and in the sense of their being useful for automatic detection of focal accents. Moreover, spectral emphasis turns out to be the more reliable correlate, as the influence on it of position in the phrase, word accent and vowel height was less pronounced and as it proved a better predictor of focal accents in general and for a majority of the speakers. Finally, the study has resulted in data for overall intensity and spectral emphasis that might prove important in modeling for speech synthesis.

  • 32.
    Heldner, Mattias
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik, Avdelningen för fonetik.
    Proceedings from FONETIK 2014: Stockholm, June 9-11, 20142014Proceedings (redaktörskap) (Övrigt vetenskapligt)
    Ladda ner fulltext (pdf)
    Proceedings from FONETIK 2014
  • 33.
    Heldner, Mattias
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik, Avdelningen för fonetik.
    Proceedings from FONETIK 2019 Stockholm, June 10–12, 20192019Proceedings (redaktörskap) (Övrigt vetenskapligt)
  • 34.
    Heldner, Mattias
    Umeå University, Department of Philosophy and Linguistics.
    Spectral emphasis as a perceptual cue to prominence2001Ingår i: TMH-QPSR 42, Stockholm: KTH , 2001, s. 51-57Kapitel i bok, del av antologi (Övrigt vetenskapligt)
  • 35.
    Heldner, Mattias
    Umeå University, Department of Philosophy and Linguistics.
    Spectral emphasis as an additional source of information in accent detection2001Ingår i: Prosody 2001: ISCA Tutorial and Research Workshop on Prosody in Speech Recognition and Understanding, Red Bank, NJ, USA: ISCA , 2001, s. 57-60Kapitel i bok, del av antologi (Refereegranskat)
  • 36.
    Heldner, Mattias
    Umeå University, Department of Philosophy and Linguistics.
    To what extent is perceived focus determined by F0-cues?1997Ingår i: Eurospeech 97, Proceedings / [ed] G. Kokkinakis, N. Fakotakis, E. Dermatas, Rhodes, Greece: ESCA , 1997, s. 875-877Konferensbidrag (Refereegranskat)
  • 37.
    Heldner, Mattias
    et al.
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik, Avdelningen för fonetik.
    Carlsson, Denise
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik, Avdelningen för fonetik.
    Wlodarczak, Marcin
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik, Avdelningen för fonetik.
    Does lung volume size affect respiratory rate and utterance duration?2019Ingår i: Proceedings from Fonetik 2019, 2019, s. 97-102Konferensbidrag (Övrigt vetenskapligt)
    Abstract [en]

    This study explored whether lung volume size affects respiratory rate and utterance duration. The lung capacity of four women and four men was estimated with a digital spirometer. These subjects subsequently read a nonsense text aloud while their respiratory movements were registered with a Respiratory Inductance Plethysmography (RIP) system. Utterance durations were measured from the speech recordings, and respiratory cycle durations and respiratory rates were measured from the RIP recordings. This experiment did not show any relationship between lung volume size and respiratory rate or utterance duration.

    Ladda ner fulltext (pdf)
    fulltext
  • 38.
    Heldner, Mattias
    et al.
    KTH Speech, Music and Hearing.
    Edlund, Jens
    KTH Speech, Music and Hearing.
    Pauses, gaps and overlaps in conversations2010Ingår i: Journal of Phonetics, ISSN 0095-4470, E-ISSN 1095-8576, Vol. 38, nr 4, s. 555-568Artikel i tidskrift (Refereegranskat)
    Abstract [en]

    This paper explores durational aspects of pauses gaps and overlaps in three different conversational corpora with a view to challenge claims about precision timing in turn-taking Distributions of pause gap and overlap durations in conversations are presented and methodological issues regarding the statistical treatment of such distributions are discussed The results are related to published minimal response times for spoken utterances and thresholds for detection of acoustic silences in speech It is shown that turn-taking is generally less precise than is often claimed by researchers in the field of conversation analysis or interactional linguistics These results are discussed in the light of their implications for models of timing in turn-taking and for interaction control models in speech technology In particular it is argued that the proportion of speaker changes that could potentially be triggered by information immediately preceding the speaker change is large enough for reactive interaction controls models to be viable in speech technology.

  • 39.
    Heldner, Mattias
    et al.
    KTH Speech, Music and Hearing.
    Edlund, Jens
    KTH Speech, Music and Hearing.
    What turns speech into conversation?: A project description2007Ingår i: Quarterly progress and status report: proceedings from Fonetik 2007, May 30-June 1, 2007, Stockholm: Department of Speech, Music and Hearing, KTH , 2007, s. 45-48Konferensbidrag (Övrigt vetenskapligt)
  • 40.
    Heldner, Mattias
    et al.
    KTH Speech, Music and Hearing.
    Edlund, Jens
    KTH Speech, Music and Hearing.
    Carlson, Rolf
    KTH Speech, Music and Hearing.
    Interruption impossible2006Ingår i: Nordic Prosody: Proceedings of the IXth Conference, Lund 2004, Frankfurt am Main: Peter Lang Publishing Group, 2006, s. 97-105Konferensbidrag (Refereegranskat)
  • 41.
    Heldner, Mattias
    et al.
    KTH Speech, Music and Hearing.
    Edlund, Jens
    KTH Speech, Music and Hearing.
    Hirschberg, Julia
    Columbia University Computer Science.
    Pitch similarity in the vicinity of backchannels2010Ingår i: Proceedings Interspeech 2010, Makuhari, Japan: ISCA , 2010, s. 3054-3057Konferensbidrag (Refereegranskat)
  • 42.
    Heldner, Mattias
    et al.
    KTH Speech, Music and Hearing.
    Edlund, Jens
    KTH Speech, Music and Hearing.
    Hjalmarsson, Anna
    KTH Speech, Music and Hearing.
    Laskowski, Kornel
    KTH Speech, Music and Hearing.
    Very short utterances and timing in turn-taking2011Ingår i: Proceedings Interspeech 2011, Florence, Italy: ISCA , 2011, s. 2837-2840Konferensbidrag (Refereegranskat)
  • 43.
    Heldner, Mattias
    et al.
    KTH Speech, Music and Hearing.
    Edlund, Jens
    KTH Speech, Music and Hearing.
    Laskowski, Kornel
    KTH Speech, Music and Hearing.
    Pelcé, Antoine
    KTH Speech, Music and Hearing.
    Prosodic features in the vicinity of silences and overlaps2009Ingår i: Nordic Prosody: Proceedings of the Xth Conference, Helsinki 2008, Frankfurt am Main: Peter Lang , 2009, s. 95-105Konferensbidrag (Refereegranskat)
  • 44.
    Heldner, Mattias
    et al.
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik, Avdelningen för fonetik.
    Hjalmarsson, Anna
    Edlund, Jens
    Backchannel relevance spaces2013Ingår i: Nordic Prosody: Proceedings of XIth Conference, Tartu 2012 / [ed] Eva Liina Asu, Pärtel Lippus, Frankfurt am Main, Germany: Peter Lang Publishing Group, 2013, s. 137-146Konferensbidrag (Refereegranskat)
    Abstract [en]

    This contribution introduces backchannel relevance spaces – intervals where it is relevant for a listener in a conversation to produce a backchannel. By annotating and comparing actual visual and vocal backchannels with potential backchannels established using a group of subjects acting as third-party listeners, we show (i) that visual only backchannels represent a substantial proportion of all backchannels; and (ii) that there are more opportunities for backchannels (i.e. potential backchannels or backchannel relevance spaces) than there are actual vocal and visual backchannels. These findings indicate that backchannel relevance spaces enable more accurate acoustic, prosodic, lexical (et cetera) descriptions of backchannel inviting cues than descriptions based on the context of actual vocal backchannels only.

    Ladda ner fulltext (pdf)
    Backchannel relevance spaces
  • 45.
    Heldner, Mattias
    et al.
    KTH Speech, Music and Hearing.
    Megyesi, Beáta
    KTH Speech, Music and Hearing.
    Exploring the prosody-syntax interface in conversations2003Ingår i: ceedings ICPhS 2003, Barcelona, Spain: ICPhS , 2003, s. 2501-2504Konferensbidrag (Refereegranskat)
  • 46.
    Heldner, Mattias
    et al.
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik, Avdelningen för fonetik.
    Riad, Tomas
    Stockholms universitet, Humanistiska fakulteten, Institutionen för svenska och flerspråkighet, Svenska/Nordiska språk. Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik, Avdelningen för fonetik.
    Sundberg, Johan
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik.
    Włodarczak, Marcin
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik, Avdelningen för fonetik.
    Zora, Hatice
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik, Avdelningen för fonetik.
    Pride and prominence2021Ingår i: Working Papers in Linguistics: Proceedings of Fonetik 2021, 2021, s. 1-6Konferensbidrag (Övrigt vetenskapligt)
    Abstract [en]

    Given the importance of the entire voice source in prominence expression, this paper aims to explore whether the word accent distinction can be defined by the voice quality dynamics moving beyond the tonal movements.To this end, a list of word accent pairs in Central Swedish were recorded and analysed based on a set of acoustic features extracted from the accelerometer signal. The results indicate that the tonal movements are indeed accompanied by the voice quality dynamics such as intensity, periodicity, harmonic richness and spectral tilt, and suggest that these parameters might contribute to the perception of one vs. two peaks associated with the word accent distinction in this regional variant of Swedish. These results, although based on limited data, are of crucial importance for the designation of voice quality variation as a prosodic feature per se.

  • 47.
    Heldner, Mattias
    et al.
    KTH Speech, Music and Hearing.
    Strangert, Eva
    Umeå University.
    Temporal effects of focus in Swedish2001Ingår i: Journal of Phonetics, ISSN 0095-4470, E-ISSN 1095-8576, Vol. 29, nr 3, s. 329-361Artikel i tidskrift (Refereegranskat)
    Abstract [en]

    The four experiments reported concern the amount and domain of lengthening associated with focal accents in Swedish. Word, syllable and segment durations were measured in read sentences with focus in different positions. As expected, words with focal accents were longer than nonfocal words in general, but the amount of lengthening varied greatly, primarily due to speaker differences but also to position in the phrase and the word accent distinction. Most of the lengthening occurred within the stressed syllable. An analysis of the internal structure of stressed syllables showed that the phonologically long segments-whether vowels or consonants-were lengthened most, while the phonologically short vowels were hardly affected at all. Through this nonlinear lengthening, the contrast between long and short vowels in stressed syllables was sharpened in focus. Thus, the domain of focal accent lengthening includes at least the stressed syllable. Also, an unstressed syllable immediately to the right of the stressed one was lengthened in focus, while initial unstressed syllables, as well as unstressed syllables to the right of the first unstressed one, were not lengthened. Thus, we assume the domain of focal accent lengthening in Swedish to be restricted to the stressed syllable and the immediately following unstressed one.

  • 48.
    Heldner, Mattias
    et al.
    Umeå University, Department of Philosophy and Linguistics.
    Strangert, Eva
    Umeå University, Department of Philosophy and Linguistics.
    Deschamps, Thierry
    Umeå University, Department of Philosophy and Linguistics.
    A focus detector using overall intensity and high frequency emphasis1999Ingår i: Proceedings of the XIVth International Congress of Phonetic Sciences: San Francisco, 1-7 August 1999 / [ed] John J. Ohala, Berkeley, Calif: Linguistics department, Univ. of California , 1999, s. 1491-1493Konferensbidrag (Refereegranskat)
  • 49.
    Heldner, Mattias
    et al.
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik, Avdelningen för fonetik.
    Wagner, Petra
    Włodarczak, Marcin
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik, Avdelningen för fonetik.
    Deep throat as a source of information2018Ingår i: Proceedings Fonetik 2018 / [ed] Åsa Abelin, Yasuko Nagano-Madsen, Gothenburg: University of Gothenburg, 2018, s. 33-38Konferensbidrag (Övrigt vetenskapligt)
    Abstract [en]

    In this pilot study we explore the signal from an accelerometer placed on the tracheal wall (below the glottis) for obtaining robust voice quality estimates. We investigate cepstral peak prominence smooth, H1-H2 and alpha ratio for distinguishing between breathy, modal and pressed phonation across six (sustained) vowel qualities produced by four speakers and including a systematic variation of pitch. We show that throat signal spectra are unaffected by vocal tract resonances, F0 and speaker variation while retaining sensitivity to voice quality dynamics. We conclude that the throat signal is a promising tool for studying communicative functions of voice prosody in speech communication.

    Ladda ner fulltext (pdf)
    fulltext
  • 50.
    Heldner, Mattias
    et al.
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik, Avdelningen för fonetik.
    Wlodarczak, Marcin
    Stockholms universitet, Humanistiska fakulteten, Institutionen för lingvistik, Avdelningen för fonetik.
    Beňuš, Štefan
    Gravano, Agustín
    Voice Quality as a Turn-Taking Cue2019Ingår i: Proceedings of Interspeech 2019 / [ed] Gernot Kubin, Zdravko Kačič, The International Speech Communication Association (ISCA), 2019, s. 4165-4169Konferensbidrag (Refereegranskat)
    Abstract [en]

    This work revisits the idea that voice quality dynamics (VQ) contributes to conveying pragmatic distinctions, with two case studies to further test this idea. First, we explore VQ as a turn-taking cue, and then as a cue for distinguishing between different functions of affirmative cue words. We employ acoustic VQ measures claimed to be better suited for continuous speech than those in own previous work. Both cases indicate that the degree of periodicity (as measured by CPPS) is indeed relevant in the production of the different pragmatic functions. In particular, turn-yielding is characterized by lower periodicity, sometimes accompanied by presence of creaky voice. Periodicity also distinguishes between backchannels, agreements and acknowledgements.

    Ladda ner fulltext (pdf)
    fulltext
12 1 - 50 av 97
RefereraExporteraLänk till träfflistan
Permanent länk
Referera
Referensformat
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Annat format
Fler format
Språk
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Annat språk
Fler språk
Utmatningsformat
  • html
  • text
  • asciidoc
  • rtf