Ändra sökning
Avgränsa sökresultatet
45678910 301 - 350 av 692
RefereraExporteraLänk till träfflistan
Permanent länk
Referera
Referensformat
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Annat format
Fler format
Språk
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Annat språk
Fler språk
Utmatningsformat
  • html
  • text
  • asciidoc
  • rtf
Träffar per sida
  • 5
  • 10
  • 20
  • 50
  • 100
  • 250
Sortering
  • Standard (Relevans)
  • Författare A-Ö
  • Författare Ö-A
  • Titel A-Ö
  • Titel Ö-A
  • Publikationstyp A-Ö
  • Publikationstyp Ö-A
  • Äldst först
  • Nyast först
  • Skapad (Äldst först)
  • Skapad (Nyast först)
  • Senast uppdaterad (Äldst först)
  • Senast uppdaterad (Nyast först)
  • Disputationsdatum (tidigaste först)
  • Disputationsdatum (senaste först)
  • Standard (Relevans)
  • Författare A-Ö
  • Författare Ö-A
  • Titel A-Ö
  • Titel Ö-A
  • Publikationstyp A-Ö
  • Publikationstyp Ö-A
  • Äldst först
  • Nyast först
  • Skapad (Äldst först)
  • Skapad (Nyast först)
  • Senast uppdaterad (Äldst först)
  • Senast uppdaterad (Nyast först)
  • Disputationsdatum (tidigaste först)
  • Disputationsdatum (senaste först)
Markera
Maxantalet träffar du kan exportera från sökgränssnittet är 250. Vid större uttag använd dig av utsökningar.
  • 301.
    Hansen, Kjetil Falkenberg
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Medieteknik och interaktionsdesign, MID. KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Musikakustik.
    Thordarson, Hilmar
    Karlsson, Haraldur
    How interactive are interactive installations?: How musical are musical interfaces? Testing interactivity and playability in students' projects2006Ingår i: Proceedings of NoMute / [ed] Tro, Jan, Trondheim, Norway, 2006, s. 23-27Konferensbidrag (Refereegranskat)
    Abstract [en]

    This paper describes the results and experiences from an experiment at The Tone and Video Lab at Listaháskóli Íslands sponsored by ConGAS, European cost action 287. Emphasis was to look at interactive interfaces, and the research aimed at finding a correspondence between how the interaction with an art piece looked, its potential for allowing interaction and expressive communication, and its appeal. Questionnaires were used to study aspects such as expressivity and playability. Most artists have a quite good appreciation on how their piece will be experienced. The results show some tendencies with regards to how easy it is to interact compared to the potential for practising skilled performances. How attractive the interaction looks seems to be of less importance.

  • 302.
    Hansson, Sven Ove
    KTH, Skolan för arkitektur och samhällsbyggnad (ABE), Filosofi och teknikhistoria, Filosofi.
    Blockage Revision2015Ingår i: Journal of Logic, Language and Information, ISSN 0925-8531, E-ISSN 1572-9583Artikel i tidskrift (Refereegranskat)
    Abstract [en]

    Blockage revision is a version of descriptor revision, i.e. belief change in which a belief set K is changed with inputs whose success conditions are metalinguistic expressions containing the belief predicate (Formula presented.). This is a highly general framework that allows a single revision operator (Formula presented.) to take inputs corresponding to sentential revision ((Formula presented.)), contraction ((Formula presented.)) as well as more complex and composite operations. In blockage revision, such an operation is based on a relation (Formula presented.) of blockage among the set of potential outcomes. (Formula presented.) signifies that if X satisfies the success condition of a belief change, then Y cannot be its outcome. The properties of blockage revision are investigated, and conditions on the blocking relation are specified that characterize various properties of the resulting operation of change.

  • 303. Havel, Miriam
    et al.
    Hofmann, Gert
    Muerbe, Dirk
    Sundberg, Johan
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Musikakustik. Univ Coll Mus Educ Stockholm, Stockholm, Sweden.
    Contribution of Paranasal Sinuses to the Acoustic Properties of the Nasal Tract2014Ingår i: Folia Phoniatrica et Logopaedica, ISSN 1021-7762, E-ISSN 1421-9972, Vol. 66, nr 3, s. 109-114Artikel i tidskrift (Refereegranskat)
    Abstract [en]

    Background: The contribution of the nasal and paranasal cavities to the vocal tract resonator properties is unclear. Here we investigate these resonance phenomena of the sinonasal tract in isolation in a cadaver and compare the results with those gained in a simplified brass tube model. Methods: The resonance characteristics were measured as the response to sine sweep excitation from an earphone. In the brass model the earphone was placed at the closed end and in the cadaver in the epipharynx. The response was picked up by a microphone placed at the open end of the model and at the nostrils, respectively. A shunting cavity with varied volumes was connected to the model and the effects on the response curve were determined. In the cadaver, different conditions with blocked and unblocked middle meatus and sphenoidal ostium were tested. Additionally, infundibulotomy was performed allowing direct access to and selective occlusion of the maxillary ostium. Results: In both the brass model and the cadaver, a baseline condition with no cavities included produced response curves with clear resonance peaks separated by valleys. Marked dips occurred when shunting cavities were attached to the model. The frequencies of these dips decreased with increasing shunting volume. In the cadaver, a marked dip was observed after removing the unilateral occlusion of the middle meatus and the sphenoidal ostium. Another marked dip was detected at low frequency after removal of the occlusion of the maxillary ostium following infundibulotomy. Conclusion: Combining measurements on a simplified nasal model with measurements in a cadaveric sinonasal tract seems a promising method for shedding light on the acoustic properties of the nasal resonator.

  • 304. Heldner, Mattias
    et al.
    Edlund, Jens
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Continuer relevance spaces2012Ingår i: Proc. of Nordic Prosody XI, Tartu, Estonia, 2012Konferensbidrag (Övrigt vetenskapligt)
  • 305.
    Heldner, Mattias
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Edlund, Jens
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Prosodic cues for interaction control in spoken dialogue systems2006Ingår i: Proceedings of Fonetik 2006, Lund, Sweden: Lund University, Centre for Languages & Literature, Dept. of Linguistics & Phonetics , 2006, s. 53-56Konferensbidrag (Övrigt vetenskapligt)
    Abstract [en]

    This paper discusses the feasibility of using prosodic features for interaction control in spoken dialogue systems, and points to experimental evidence that automatically extracted prosodic features can be used to improve the efficiency of identifying relevant places at which a machine can legitimately begin to talk to a human interlocutor, as well as to shorten system response times.

  • 306.
    Heldner, Mattias
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Edlund, Jens
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    What turns speech into conversation?: A project description2007Ingår i: TMH-QPSR, ISSN 1104-5787, Vol. 50, nr 1, s. 45-48Artikel i tidskrift (Refereegranskat)
    Abstract [en]

    The project Vad gör tal till samtal? (What turns speech into conversation?) takes as its starting point that while conversation must be considered the primary kind of speech, we are still far better at modelling monologue than dialogue, in theory as well as for speech technology applications. There are also good reasons to assume that conversation contains a number of features that are not found in other kinds of speech, including, among other things, the active cooperation among interlocutors to control the interaction, and to establish common ground. Through this project, we hope to improve the situation by investigating features that are specific to human-human conversation – features that turns speech into conversation. We will focus on acoustic and prosodic aspects of such features.

  • 307.
    Heldner, Mattias
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Edlund, Jens
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Björkenstam, Tomas
    Automatically extracted F0 features as acoustic correlates of prosodic boundaries2004Ingår i: Fonetik 2004: Proc of The XVIIth Swedish Phonetics Conference, Stockholm University, 2004, s. 52-55Konferensbidrag (Refereegranskat)
    Abstract [en]

    This work presents preliminary results of an investigation of various automatically extracted F0 features as acoustic correlates of prosodic boundaries. The F0 features were primarily intended to capture phenomena such as boundary tones, F0 resets across boundaries and position in the speaker's F0 range. While there were no correspondences between boundary tones and boundaries, the reset and range features appeared to separate boundaries from no boundaries fairly well.

  • 308.
    Heldner, Mattias
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Edlund, Jens
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Carlson, Rolf
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Interruption impossible2006Ingår i: Nordic Prosody: Proceedings of the IXth Conference, Lund 2004 / [ed] Bruce, G.; Horne, M., Frankfurt am Main, Germany, 2006, s. 97-105Konferensbidrag (Refereegranskat)
    Abstract [en]

    Most current work on spoken human-computer interaction has so far concentrated on interactions between a single user and a dialogue system. The advent of ideas of the computer or dialogue system as a conversational partner in a group of humans, for example within the CHIL-project1 and elsewhere (e.g. Kirchhoff & Ostendorf, 2003), introduces new requirements on the capabilities of the dialogue system. Among other things, the computer as a participant in a multi-part conversation has to appreciate the human turn-taking system, in order to time its' own interjections appropriately. As the role of a conversational computer is likely to be to support human collaboration, rather than to guide or control it, it is particularly important that it does not interrupt or disturb the human participants. The ultimate goal of the work presented here is to predict suitable places for turn-takings, as well as positions where it is impossible for a conversational computer to interrupt without irritating the human interlocutors.

  • 309.
    Heldner, Mattias
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Edlund, Jens
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Hirschberg, Julia
    Pitch similarity in the vicinity of backchannels2010Ingår i: Proceedings of the 11th Annual Conference of the International Speech Communication Association, INTERSPEECH 2010, Makuhari, Japan, 2010, s. 3054-3057Konferensbidrag (Refereegranskat)
    Abstract [en]

    Dynamic modeling of spoken dialogue seeks to capture how interlocutors change their speech over the course of a conversation. Much work has focused on how speakers adapt or entrain to different aspects of one another’s speaking style. In this paper we focus on local aspects of this adaptation. We investigate the relationship between backchannels and the interlocutor utterances that precede them with respect to pitch. We demonstrate that the pitch of backchannels is more similar to the immediately preceding utterance than nonbackchannels. This inter-speaker pitch relationship captures the same distinctions as more cumbersome intra-speaker relations, and supports the intuition that, in terms of pitch, such similarity may be one of the mechanisms by which backchannels are rendered ’unobtrusive’.

  • 310.
    Heldner, Mattias
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Edlund, Jens
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Hjalmarsson, Anna
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Laskowski, Kornel
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Very short utterances and timing in turn-taking2011Ingår i: Proceedings of Interspeech 2011, 2011, s. 2848-2851Konferensbidrag (Refereegranskat)
    Abstract [en]

    This work explores the timing of very short utterances in conversations, as well as the effects of excluding intervals adjacent to such utterances from distributions of between-speaker interval durations. The results show that very short utterances are more precisely timed to the preceding utterance than longer utterances in terms of a smaller variance and a larger proportion of no-gap-no-overlaps. Excluding intervals adjacent to very short utterances furthermore results in measures of central tendency closer to zero (i.e. no-gap-no-overlaps) as well as larger variance (i.e. relatively longer gaps and overlaps).

  • 311.
    Heldner, Mattias
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Edlund, Jens
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Laskowski, Kornel
    Pelcé, Antoine
    Prosodic features in the vicinity of pauses, gaps and overlaps2009Ingår i: Nordic Prosody: Proceedings of the Xth Conference / [ed] Vainio, Martti; Aulanko, Reijo; Aaltonen, Olli, Frankfurt am Main: Peter Lang , 2009, s. 95-106Konferensbidrag (Refereegranskat)
    Abstract [en]

    In this study, we describe the range of prosodic variation observed in two types of dialogue contexts, using fully automatic methods. The first type of context is that of speaker-changes, or transitions from only one participant speaking to only the other, involving either acoustic silences or acoustic overlaps. The second type of context is comprised of mutual silences or overlaps where a speaker change could in principle occur but does not. For lack of a better term, we will refer to these contexts as non-speaker-changes. More specifically, we investigate F0 patterns in the intervals immediately preceding overlaps and silences – in order to assess whether prosody before overlaps or silences may invite or inhibit speaker change.

  • 312.
    Heldner, Mattias
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Hjalmarsson, Anna
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Edlund, Jens
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Backchannel relevance spaces2013Ingår i: Prosody: Proceedings of the XIth Conference, Tartu 2012 / [ed] Eva Liina / Lippus, Pärtel, Peter Lang Publishing Group, 2013, s. 137-146Konferensbidrag (Refereegranskat)
  • 313.
    Henriksson, Aron
    et al.
    Stockholms universitet, Institutionen för data- och systemvetenskap.
    Zhao, Jing
    Stockholms universitet, Institutionen för data- och systemvetenskap.
    Boström, Henrik
    Stockholms universitet, Institutionen för data- och systemvetenskap.
    Dalianis, Hercules
    Stockholms universitet, Institutionen för data- och systemvetenskap.
    Modeling Electronic Health Records in Ensembles of Semantic Spaces for Adverse Drug Event Detection2015Ingår i: 2015 IEEE International Conference on Bioinformatics and Biomedicine: Proceedings / [ed] Jun (Luke) Huan et al., IEEE Computer Society, 2015, s. 343-350, artikel-id 7359705Konferensbidrag (Refereegranskat)
    Abstract [en]

    Electronic health records (EHRs) are emerging as a potentially valuable source for pharmacovigilance; however, adverse drug events (ADEs), which can be encoded in EHRs by a set of diagnosis codes, are heavily underreported. Alerting systems, able to detect potential ADEs on the basis of patient- specific EHR data, would help to mitigate this problem. To that end, the use of machine learning has proven to be both efficient and effective; however, challenges remain in representing the heterogeneous EHR data, which moreover tends to be high- dimensional and exceedingly sparse, in a manner conducive to learning high-performing predictive models. Prior work has shown that distributional semantics – that is, natural language processing methods that, traditionally, model the meaning of words in semantic (vector) space on the basis of co-occurrence information – can be exploited to create effective representations of sequential EHR data, not only free-text in clinical notes but also various clinical events such as diagnoses, drugs and measurements. When modeling data in semantic space, an im- portant design decision concerns the size of the context window around an object of interest, which governs the scope of co- occurrence information that is taken into account and affects the composition of the resulting semantic space. Here, we report on experiments conducted on 27 clinical datasets, demonstrating that performance can be significantly improved by modeling EHR data in ensembles of semantic spaces, consisting of multiple semantic spaces built with different context window sizes. A follow-up investigation is conducted to study the impact on predictive performance as increasingly more semantic spaces are included in the ensemble, demonstrating that accuracy tends to improve with the number of semantic spaces, albeit not monotonically so. Finally, a number of different strategies for combining the semantic spaces are explored, demonstrating the advantage of early (feature) fusion over late (classifier) fusion. Ensembles of semantic spaces allow multiple views of (sparse) data to be captured (densely) and thereby enable improved performance to be obtained on the task of detecting ADEs in EHRs.

  • 314.
    Henriksson, Aron
    et al.
    Stockholms universitet, Institutionen för data- och systemvetenskap.
    Zhao, Jing
    Stockholms universitet, Institutionen för data- och systemvetenskap.
    Boström, Henrik
    Stockholms universitet, Institutionen för data- och systemvetenskap.
    Dalianis, Hercules
    Stockholms universitet, Institutionen för data- och systemvetenskap.
    Modeling Heterogeneous Clinical Sequence Data in Semantic Space for Adverse Drug Event Detection2015Ingår i: Proceedings of the 2015 IEEE International Conference on Data Science and Advanced Analytics / [ed] Eric Gaussier, Longbing Cao, Patrick Gallinari, James Kwok, Gabriella Pasi, Osmar Zaiane, Institute of Electrical and Electronics Engineers (IEEE), 2015, s. 792-799, artikel-id 7344867Konferensbidrag (Refereegranskat)
    Abstract [en]

    The enormous amounts of data that are continuously recorded in electronic health record systems offer ample opportunities for data science applications to improve healthcare. There are, however, challenges involved in using such data for machine learning, such as high dimensionality and sparsity, as well as an inherent heterogeneity that does not allow the distinct types of clinical data to be treated in an identical manner. On the other hand, there are also similarities across data types that may be exploited, e.g., the possibility of representing some of them as sequences. Here, we apply the notions underlying distributional semantics, i.e., methods that model the meaning of words in semantic (vector) space on the basis of co-occurrence information, to four distinct types of clinical data: free-text notes, on the one hand, and clinical events, in the form of diagnosis codes, drug codes and measurements, on the other hand. Each semantic space contains continuous vector representations for every unique word and event, which can then be used to create representations of, e.g., care episodes that, in turn, can be exploited by the learning algorithm. This approach does not only reduce sparsity, but also takes into account, and explicitly models, similarities between various items, and it does so in an entirely data-driven fashion. Here, we report on a series of experiments using the random forest learning algorithm that demonstrate the effectiveness, in terms of accuracy and area under ROC curve, of the proposed representation form over the commonly used bag-of-items counterpart. The experiments are conducted on 27 real datasets that each involves the (binary) classification task of detecting a particular adverse drug event. It is also shown that combining structured and unstructured data leads to significant improvements over using only one of them.

  • 315.
    Hincks, Rebecca
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH.
    Computer support for learners of spoken English2005Doktorsavhandling, monografi (Övrigt vetenskapligt)
    Abstract [en]

    This thesis concerns the use of speech technology to support the process of learning the English language. It applies theories of computer-assisted language learning and second language acquisition to address the needs of beginning, intermediate and advanced students of English for specific purposes.

    The thesis includes an evaluation of speech-recognition-based pronunciation software, based on a controlled study of a group of immigrant engineers. The study finds that while the weaker students may have benefited from their software practice, the pronun¬ciation ability of the better students did not improve.

    The linguistic needs of advanced and intermediate Swedish-native students of English are addressed in a study using multimodal speech synthesis in an interactive exercise demonstrating differences in the placement of lexical stress in two Swedish-English cognates. A speech database consisting of 28 ten-minute oral presentations made by these learners is described, and an analysis of pronunciation errors is pre¬sented. Eighteen of the presentations are further analyzed with regard to the normalized standard deviation of fundamental frequency over 10-second long samples of speech, termed pitch variation quotient (PVQ). The PVQ is found to range from 6% to 34% in samples of speech, with mean levels of PVQ per presentation ranging from 11% to 24%. Males are found to use more pitch variation than females. Females who are more proficient in English use more pitch variation than the less profi¬cient females. A perceptual experiment tests the relationship between PVQ and impressions of speaker liveliness. An overall correlation of .83 is found. Temporal variables in the presentation speech are also studied.

    A bilingual database where five speakers make the same presentation in both English and Swedish is studied to examine effects of using a second language on presentation prosody. Little intra-speaker difference in pitch variation is found, but these speakers speak on average 20% faster when using their native language. The thesis concludes with a discussion of how the results could be applied in a proposed feedback mechanism for practicing and assessing oral presentations, concept¬ualized as a ‘speech checker.’ Potential users of the system would include native as well as non-native speakers of English.

  • 316.
    Hincks, Rebecca
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH.
    Measures and perceptions of liveliness in student presentation speech: A proposal for an automatic feedback mechanism2005Ingår i: Systeme, ISSN 1022-9280, Vol. 33, nr 4, s. 575-591Artikel i tidskrift (Refereegranskat)
    Abstract [en]

    This paper analyzes prosodic variables in a corpus of eighteen oral presentations made by students of Technical English, all of whom were native speakers of Swedish. The focus is on the extent to which speakers were able to use their voices in a lively manner, and the hypothesis tested is that speakers who had high pitch variation as they spoke would be perceived as livelier speakers. A metric (termed PVQ), derived from the standard deviation in fundamental frequency, is proposed as a measure of pitch variation. Composite listener ratings of liveliness for nine 10-s samples of speech per speaker correlate strongly (r = .83, n = 18, p < .01) with the PVQ metric. Liveliness ratings for individual 10-s samples of speech show moderate but significant (n = 81, p < .01) correlations: r = .70 for males and r = .64 for females. The paper also investigates rate of speech and fluency variables in this corpus of L2 English. An application for this research is in presentation skills training, where computer feedback could be provided for speaking rate and the extent to which speakers have been able to use their voices in an engaging manner.

  • 317.
    Hincks, Rebecca
    KTH, Skolan för teknikvetenskaplig kommunikation och lärande (ECE), Lärande, Språk och kommunikation.
    Measuring liveliness in presentation speech2005Ingår i: Proceedings of Interspeech 2005, Lisbon, 2005, s. 765-768Konferensbidrag (Refereegranskat)
    Abstract [en]

    This paper proposes that speech analysis be used to quantifyprosodic variables in presentation speech, and reports theresults of a perception test of speaker liveliness. The test materialwas taken from a corpus of oral presentations made by18 Swedish native students of Technical English. Livelinessratings from a panel of eight judges correlated strongly withnormalized standard deviation of F0 and, for female speakers,with mean length of runs, which is the number of syllablesbetween pauses of >250 ms. An application of these findingswould be in the development of a feedback mechanism for theprosody of public speaking.

  • 318.
    Hincks, Rebecca
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Språk och kommunikation (stängd 2011-01-01).
    Presenting in English and Swedish2005Ingår i: Proceedings of Fonetik 2005, Göteborg, 2005, s. 45-48Konferensbidrag (Övrigt vetenskapligt)
    Abstract [en]

    This paper reports on a comparison of prosodicvariables from oral presentations in a first andsecond language. Five Swedish natives whospeak English at the advanced-intermediatelevel were recorded as they made the samepresentation twice, once in English and once inSwedish. Though it was expected that speakerswould use more pitch variation when theyspoke Swedish, three of the five speakersshowed no significant difference between thetwo languages. All speakers spoke more quicklyin Swedish, the mean being 20% faster.

  • 319.
    Hincks, Rebecca
    KTH, Skolan för teknikvetenskaplig kommunikation och lärande (ECE), Lärande, Språk och kommunikation.
    Presenting in English or Swedish: Differences in speaking rate2008Ingår i: Proceedings of Fonetik 2008 / [ed] Eriksson, A.; Lindh, J., Gothenburg: Gothenburg University Department of Linguistics, 2008, s. 21-24Konferensbidrag (Refereegranskat)
    Abstract [en]

    This paper attempts to quantify differences in speaking rates in first and second languages, in the context of the growth of English as a lingua franca, where more L2 speakers than ever be-fore are using English to perform tasks in their working environments. One such task is the oral presentation. The subjects in this study were fourteen fluent English second language speakers who held the same oral presentation twice, once in English and once in their native Swedish. The temporal variables of phrase length (mean length of runs in syllables) and speaking rate in syllables per second were cal-culated for each language. Speaking rate was found to be 23% slower when using the second language, and phrase length was found to be 24% shorter.

  • 320.
    Hincks, Rebecca
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Språk och kommunikation (stängd 2011-01-01).
    Processing the prosody of oral presentations2004Ingår i: Proc InSTIL/ICALL2004 NLP and Speech Technologies in Advanced Language Learning / [ed] Delmonte, R.; Delcloque, P.; Tonellli, S., Venice, Italy, 2004, s. 63-66Konferensbidrag (Refereegranskat)
    Abstract [en]

    Standard advice to people preparing to speak in public is to use a “lively” voice. A lively voice is described as one that varies in intonation, rhythm and loudness: qualities that can be analyzed using speech analysis software. This paper reports on a study analyzing pitch variation as a measure of speaker liveliness. A potential application of this approach for analysis would be for rehearsing or assessing the prosody of oral presentations. While public speaking can be intimidating even to native speakers, second language users are especially challenged, particularly when it comes to using their voices in a prosodically engaging manner.The material is a database of audio recordings of twenty 10-minute student oral presentations, where all speakers were college-age Swedes studying Technical English. The speech has been processed using the analysis software WaveSurfer for pitch extraction. Speaker liveliness has been measured as the standard deviation from the mean fundamental frequency over 10-second periods of speech. The standard deviations have been normal¬ized (by division with the mean frequency) to obtain a value termed the pitch dynamism quotient (PDQ). Mean values (for ten minutes of speech) of PDQ per speaker range from a low of 0.11 to a high of 0.235. Individual values for 10-second segments range from lows of 0.06 to highs of 0.36.

  • 321. Hincks, Rebecca
    Standard deviation of F0 in student monologue2004Ingår i: Proc of The XVIIth Swedish Phonetics Conference, Fonetik 2004, Stockholm University, 2004, s. 132-135Konferensbidrag (Övrigt vetenskapligt)
    Abstract [en]

    Twenty ten-minute oral presentations made by Swedish students speaking English have been analyzed with respect to the standard deviation of F0 over long stretches of speech. Values have been normalized by division with the mean. Results show a strong correlation between pro-ficiency in English and pitch variation for male speakers but not for females. The results also identify monotone and disfluent speakers.

  • 322.
    Hincks, Rebecca
    KTH, Skolan för teknikvetenskaplig kommunikation och lärande (ECE), Avdelningen för bibliotek, språk och ARC, Språk och kommunikation.
    Suprasegmentals: Stress2013Ingår i: The Encyclopedia of Applied Linguistics / [ed] Chapelle, C., Wiley Blackwell , 2013Kapitel i bok, del av antologi (Refereegranskat)
    Abstract [en]

    Stress is a feature of pronunciation in which a syllable is given more emphasis than surrounding syllables.

  • 323.
    Hincks, Rebecca
    KTH, Skolan för teknikvetenskaplig kommunikation och lärande (ECE), Avdelningen för bibliotek, språk och ARC, Språk och kommunikation.
    Technology and Phonetics2013Ingår i: The Encyclopedia of Applied Linguistics / [ed] Chapelle, C., Wiley Blackwell , 2013Kapitel i bok, del av antologi (Refereegranskat)
    Abstract [en]

    Sound cannot be studied methodically unless it can be captured in some way, and so from the phonograph to the tape recorder to the computer, the development of new technologies has facilitated the study of phonetics. Indeed, it is difficult to imagine studying phonetics without technology. Other entries in this encyclopedia discuss the topics of acoustic phonetics, speech analysis software, automatic speech recognition, text-to-speech synthesis, and computer-assisted pronunciation teaching. This entry will thus simply provide an overview of the position of technology in phonetics.

  • 324.
    Hincks, Rebecca
    et al.
    KTH, Skolan för teknikvetenskaplig kommunikation och lärande (ECE), Lärande, Språk och kommunikation.
    Edlund, Jens
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Transient visual feedback on pitch variation for Chinese speakers of English2009Ingår i: Proc. of Fonetik 2009, Stockholm, 2009Konferensbidrag (Övrigt vetenskapligt)
    Abstract [en]

    This paper reports on an experimental study comparing two groups of seven Chinese students of English who practiced oral presentations with computer feedback. Both groups imitated teacher models and could listen to recordings of their own production. The test group was also shown flashing lights that responded to the standard deviation of the fundamental frequency over the previous two seconds. The speech of the test group increased significantly more in pitch variation than the control group. These positive results suggest that this novel type of feedback could be used in training systems for speakers who have a tendency to speak in a monotone when making oral presentations.

  • 325.
    Hincks, Rebecca
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Språk och kommunikation.
    Edlund, Jens
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Using speech technology to promote increased pitch variation in oral presentations2009Ingår i: Proc. of SLaTE Workshop on Speech and Language Technology in Education, Wroxall, UK, 2009Konferensbidrag (Refereegranskat)
    Abstract [en]

    This paper reports on an experimental study comparing two groups of seven Chinese students of English who practiced oral presentations with computer feedback. Both groups imitated teacher models and could listen to recordings of their own production. The test group was also shown flashing lights that responded to the standard deviation of the fundamental frequency over the previous two seconds. The speech of the test group increased significantly more in pitch variation than the control group. These positive results suggest that this novel type of feedback could be used in training systems for speakers who have a tendency to speak in a monotone when making oral presentations.

  • 326. Hiraga, Rumi
    et al.
    Hansen, Kjetil Falkenberg
    KTH, Skolan för datavetenskap och kommunikation (CSC), Medieteknik och interaktionsdesign, MID. KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Musikakustik.
    Sound Preferences of Persons with Hearing Loss Playing an Audio-Based Computer Game2013Ingår i: Proceedings of the 3rd ACM international workshop on Interactive multimedia on mobile & portable devices, New York: ACM Digital Library, 2013, s. 25-30Konferensbidrag (Refereegranskat)
    Abstract [en]

    We performed an experiment to investigate differences between persons with and without hearing losses when playing a novel audio-based game on a tablet computer, and how persons with hearing losses appreciated the game when they played it with three different types of sound material - speech, music, or mixed speech and music. We analyzed game log files and participants' self-assessments and obtained results showing that there were significant differences between the two participant groups in terms of whether they completed the game. Moreover, the hearing loss group showed a preference for music among the three types of sounds and for the game itself. The two groups listened to music in different ways: hearing participants worked with the music material differently compared with other two types of material, implying that music is the most difficult among the three types. The hearing loss group showed preference for the music only-condition, which is consistent with the results from preliminary experiments we have done. We suggest that this novel game has the potential to improve the listening ability of persons with a hearing loss.

  • 327. Hiraga, Rumi
    et al.
    Hansen, Kjetil Falkenberg
    KTH, Skolan för datavetenskap och kommunikation (CSC), Medieteknik och interaktionsdesign, MID.
    Kano, Naoya
    Matsubara, Masaki
    Terasawa, Hiroko
    Tabuchi, Keiji
    Music perception of hearing-impaired persons with focus on one test subject2015Ingår i: Proceedings of 2015 IEEE International Conference on Systems, Man, and Cybernetics, IEEE , 2015Konferensbidrag (Refereegranskat)
    Abstract [en]

    We previously investigated how hearing-impaired people perceive music in several types of musical experiments. By following-up on the results of some of these experiments with a single test subject Sd, we found that the ability of the subject to perceive music was high and that she appreciated music in a way that was different from that of hearing people. In this paper, we describe three musical experiments with hearing-impaired subjects, their results, and Sd’s music perception through the experiments. The three experiments involved the Music Puzzle game, the appreciation of harmony, and tempo perception. Music Puzzle is a music game we made that is played on a tablet and is intended to be used by hearing-impaired persons as a serious game with which they can improve their hearing ability by continuously playing it. The experiment on appreciation of harmony was conducted with three subject groups, and the result showed that experience with music affected the appreciation of music accompanied with the tonal code. Tempo perception was investigated with a simple game in which the subjects tapped along with the tempo of the music. By examining the subjects’ hearing acuity in standard medical hearing tests and crossing over the results of these musical experiments, we observed that hearing acuity is not necessarily related to the perception and understanding of music.

  • 328.
    Hirschberg, Julia
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Hjalmarsson, Anna
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Elhadad, Noémie
    "You're as Sick as You Sound": Using Computational Approaches for Modeling Speaker State to Gauge Illness and Recovery2010Ingår i: Mobile Environments, Call Centers and Clinics / [ed] Neustein, Amy, Springer , 2010, s. 305-322Kapitel i bok, del av antologi (Refereegranskat)
    Abstract [en]

    Recently, researchers in computer science and engineering have begun to explore the possibility of finding speech-based correlates of various medical conditions using automatic, computational methods. If such language cues can be identified and quantified automatically, this information can be used to support diagnosis and treatment of medical conditions in clinical settings and to further fundamental research in understanding cognition. This chapter reviews computational approaches that explore communicative patterns of patients who suffer from medical conditions such as depression, autism spectrum disorders, schizophrenia, and cancer. There are two main approaches discussed: research that explores features extracted from the acoustic signal and research that focuses on lexical and semantic features. We also present some applied research that uses computational methods to develop assistive technologies. In the final sections we discuss issues related to and the future of this emerging field of research.

  • 329.
    Hjalmarsson, Anna
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    On cue - additive effects of turn-regulating phenomena in dialogue2009Ingår i: Diaholmia, 2009, s. 27-34Konferensbidrag (Refereegranskat)
    Abstract [en]

    One line of work on turn-taking in dialogue suggests that speakers react to “cues” or “signals” in the behaviour of the preceding speaker. This paper describes a perception experiment that investigates if such potential turn-taking cues affect the judgments made by non-participating listeners. The experiment was designed as a game where the task was to listen to dialogues and guess the outcome, whether there will be a speaker change or not, whenever the recording was halted. Human-human dialogues as well as dialogues where one of the human voices was replaced by a synthetic voice were used. The results show that simultaneous turn-regulating cues have a reinforcing effect on the listeners’ judgements. The more turn-holding cues, the faster the reaction time, suggesting that the subjects were more confident in their judgments. Moreover, the more cues, regardless if turn-holding or turn-yielding, the higher the agreement among subjects on the predicted outcome. For the re-synthesized voice, responses were made significantly slower; however, the judgments show that the turn-taking cues were interpreted as having similar functions as for the original human voice.

  • 330.
    Hjalmarsson, Anna
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Speaking without knowing what to say... or when to end2008Ingår i: Proceedings of SIGdial 2008, Columbus, Ohio, USA, 2008, s. 72-75Konferensbidrag (Refereegranskat)
    Abstract [en]

    Humans produce speech incrementally and on-line as the dialogue progresses using information from several different sources in parallel. A dialogue system that generates output in a stepwise manner and not in preplanned syntactically correct sentences needs to signal how new dialogue contributions relate to previous discourse. This paper describes a data collection which is the foundation for an effort towards more human-like language generation in DEAL, a spoken dialogue system developed at KTH. Two annotators labelled cue phrases in the corpus with high inter-annotator agreement (kappa coefficient 0.82).

  • 331.
    Hjalmarsson, Anna
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    The vocal intensity of turn-initial cue phrases in dialogue2010Ingår i: Proceedings of the SIGDIAL 2010 Conference: 11th Annual Meeting of the Special Interest Group onDiscourse and Dialogue, 2010, s. 225-228Konferensbidrag (Refereegranskat)
    Abstract [en]

    The present study explores the vocal intensity of turn-initial cue phrases in a corpus of dialogues in Swedish. Cue phrases convey relatively little propositional content, but have several important pragmatic functions. The majority of these entities are frequently occurring monosyllabic words such as "eh", "mm", "ja". Prosodic analysis shows that these words are produced with higher intensity than other turn-initial words are. In light of these results, it is suggested that speakers produce these expressions with high intensity in order to claim the floor. It is further shown that the difference in intensity can be measured as a dynamic inter-speaker relation over the course of a dialogue using the end of the interlocutor's previous turn as a reference point.

  • 332.
    Hjalmarsson, Anna
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Towards user modelling in conversational dialogue systems: A qualitative study of the dynamics of dialogue parameters2005Ingår i: Proceedings of Interspeech 2005, Lisbon, Portugal, 2005, s. 869-872Konferensbidrag (Refereegranskat)
    Abstract [en]

    This paper presents a qualitative study of data from a 26 subject experimental study within the multimodal, conversational dialogue system AdApt. Qualitative analysis of data is used to illustrate the dynamic variation of dialogue parameters over time. The analysis will serve as a foundation for research and future data collections in the area of adaptive dialogue systems and user modelling.

  • 333.
    Hjalmarsson, Anna
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Laskowski, Kornel
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Measuring final lengthening for speaker-change prediction2011Ingår i: Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH, Florence, Italy, 2011, s. 2076-2079Konferensbidrag (Refereegranskat)
    Abstract [en]

    We explore pre-silence syllabic lengthening as a cue for next-speakership prediction in spontaneous dialogue. When estimated using a transcription-mediated procedure, lengthening is shown to reduce error rates by 25% relative to majority class guessing. This indicates that lengthening should be exploited by dialogue systems. With that in mind, we evaluate an automatic measure of spectral envelope change, Mel-spectral flux (MSF), and show that its performance is at least as good as that of the transcription-mediated measure. Modeling MSF is likely to improve turn uptake in dialogue systems, and to benefit other applications needing an estimate of durational variability in speech.

  • 334.
    Hjalmarsson, Anna
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Oertel, Catharine
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Gaze direction as a Back-Channel inviting Cue in Dialogue2012Ingår i: IVA 2012 workshop on Realtime Conversational Virtual Agents, Santa Cruz, CA, USA, 2012Konferensbidrag (Refereegranskat)
    Abstract [en]

    In this study, we experimentally explore the relationship between gaze direction and backchannels in face-to-face interaction. The overall motivation is to use gaze direction in a virtual agent as a mean to elicit user feedback. The relationship between gaze and backchannels was tested in an experiment in which participants were asked to provide feedback when listening to a story-telling virtual agent. When speaking, the agent shifted her gaze towards the listener at predefined positions in the dialogue. The results show that listeners are more prone to backchannel when the virtual agent’s gaze is directed towards them than when it is directed away. However, there is a high response variability for different dialogue contexts which suggests that the timing of backchannels cannot be explained by gaze direction alone.

  • 335.
    Hjalmarsson, Anna
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Wik, Preben
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Brusk, J.
    Dealing with DEAL: a dialogue system for conversation training2007Ingår i: Computational linguistics - Association for Computational Linguistics (Print), ISSN 0891-2017, E-ISSN 1530-9312, s. 132-135Artikel i tidskrift (Refereegranskat)
    Abstract [en]

    We present DEAL, a spoken dialogue system for conversation training under development at KTH.DEAL is a game with a spoken language interface designed for second language learners. The system is intended as a multidisciplinary research platform where challenges and potential benefits of combining elements from computer games, dialogue systems and language learning can be explored.

  • 336.
    House, David
    KTH, Tidigare Institutioner, Tal, musik och hörsel.
    Final rises and Swedish question intonation2004Ingår i: Proc of The XVIIth Swedish Phonetics Conference, Fonetik 2004, Stockholm University, 2004, s. 56-59Konferensbidrag (Övrigt vetenskapligt)
    Abstract [en]

    Phrase-final intonation was analysed in a subcorpusof Swedish computer-directed questionutterances with the objective of investigatingthe extent to which final rises occur in spontaneousquestions, and also to see if such risesmight have pragmatic functions over and beyondthe signalling of interrogative mode. Finalrises occurred in 22 percent of the utterances.Final rises occurred mostly in conjunctionwith final focal accent. Children exhibitedthe largest percentage of final rises (32%), withwomen second (27%) and men lowest (17%).These results are viewed in relationship to resultsof related perception studies and are discussedin terms of Swedish question intonationand the pragmatic social function of rises in abiological account of intonation.

  • 337.
    House, David
    KTH, Tidigare Institutioner, Tal, musik och hörsel.
    Final rises in spontaneous Swedish computer-directed questions: incidence and function2004Ingår i: Proc of Intl Conference on Speech Prosody 2004 / [ed] Bel, B.; Marlin, I., Nara, Japan, 2004, s. 115-118Konferensbidrag (Refereegranskat)
    Abstract [en]

    Phrase-final intonation was analysed in a subcorpus of Swedish computer-directed question utterances with the objective of investigating the extent to which final rises occur in spontaneous questions, and also to see if such rises might have pragmatic functions over and beyond the signalling of interrogative mode. Final rises occurred in 22 percent of the utterances. Final rises occurred mostly in conjunction with final focal accent. Children exhibited the largest percentage of final rises (32%), with women second (27%) and men lowest (17%). These results are discussed in terms of Swedish question intonation and the pragmatic social function of rises in a biological account of intonation.

  • 338.
    House, David
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Fonetiska undersökningar av kammu2005Ingår i: Kammu - om ett folk i Laos / [ed] Lundström, H.; Svantesson, J.-O., Lund: Lunds universitetshistoriska sällskap , 2005, s. 164-167Kapitel i bok, del av antologi (Refereegranskat)
  • 339.
    House, David
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    On the interaction of audio and visual cues to friendliness in interrogative prosody2006Ingår i: Proceedings of The Nordic Conference on Multimodal Communication, 2005, Göteborg, 2006, s. 201-213Konferensbidrag (Refereegranskat)
  • 340.
    House, David
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Perception and production of phrase-final intonation in Swedish questions2006Ingår i: Nordic Prosody, Proceedings of the IXth Conference, Lund 2004 / [ed] Bruce, G.; Horne, M., Frankfurt am Main: Peter Lang , 2006, s. 127-136Konferensbidrag (Refereegranskat)
  • 341.
    House, David
    KTH, Tidigare Institutioner, Tal, musik och hörsel.
    Pitch and alignment in the perception of tone and intonation2004Ingår i: From Traditional Phonology to Modern Speech Processing / [ed] Fant, G.; Fujisaki, H.; Cao, J.; Xu, Y., Beijing: Foreign Language Teaching and Research Press , 2004, s. 189-204Kapitel i bok, del av antologi (Refereegranskat)
  • 342.
    House, David
    KTH, Tidigare Institutioner, Tal, musik och hörsel.
    Pitch and alignment in the perception of tone and intonation: pragmatic signals and biological codes2004Ingår i: Proc of International Symposium on Tonal Aspects of Languages: Emphasis on Tone Languages / [ed] Bel, B.; Marlein, I., Beijng, China, 2004, s. 93-96Konferensbidrag (Refereegranskat)
  • 343.
    House, David
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Alexanderson, Simon
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Beskow, Jonas
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    On the temporal domain of co-speech gestures: syllable, phrase or talk spurt?2015Ingår i: Proceedings of Fonetik 2015 / [ed] Lundmark Svensson, M.; Ambrazaitis, G.; van de Weijer, J., 2015, s. 63-68Konferensbidrag (Övrigt vetenskapligt)
    Abstract [en]

    This study explores the use of automatic methods to detect and extract handgesture movement co-occuring with speech. Two spontaneous dyadic dialogueswere analyzed using 3D motion-capture techniques to track hand movement.Automatic speech/non-speech detection was performed on the dialogues resultingin a series of connected talk spurts for each speaker. Temporal synchrony of onsetand offset of gesture and speech was studied between the automatic hand gesturetracking and talk spurts, and compared to an earlier study of head nods andsyllable synchronization. The results indicated onset synchronization between headnods and the syllable in the short temporal domain and between the onset of longergesture units and the talk spurt in a more extended temporal domain.

  • 344.
    House, David
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Granström, Björn
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Analyzing and modelling verbal and non-verbal communication for talking animated interface agents2007Ingår i: Fundamentals of verbal and nonverbal communication and the biometric issue / [ed] Esposito, A.; Bratanic, M.; Keller, E.; Marinaro, M., Amsterdam: IOS Press , 2007, s. 317-331Kapitel i bok, del av antologi (Refereegranskat)
  • 345.
    House, David
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Karlsson, A.
    Svantesson, J-O.
    Tayanin, D.
    On utterance-final intonation in tonal and non-tonal dialects of Kammu2009Ingår i: Proceedings of Fonetik 2009, Department of Linguistics, Stockholm University, 2009, s. 78-81Konferensbidrag (Övrigt vetenskapligt)
    Abstract [en]

    In this study we investigate utterance-final intonation in two dialects of Kammu, one tonal and one non-tonal. While the general patterns of utterance-final intonation are similar between the dialects, we do find clear evidence that the lexical tones of the tonal dialect restrict the pitch range and the realization of focus. Speaker engagement can have a strong effect on the utterance-final accent in both dialects.

  • 346.
    House, David
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Strömbergsson, Sofia
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Self-voice identification in children with phonological impairment2011Ingår i: Proceedings of the ICPhS XVII, Hong Kong, 2011, s. 886-889Konferensbidrag (Refereegranskat)
    Abstract [en]

    We report preliminary data from a study of self-voice identification in children with phonological impairment (PI), where results from 13 children with PI are compared to results from a group of children with typical speech. No difference between the two groups was found, suggesting that a phonological impairment does not affect children’s ability to recognize their recorded voices as their own. We conclude that children with PI indeed recognize their own recorded voice and that the use of recordings in therapy can be supported.

  • 347.
    Hunnicutt, Sharon
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Nozadze, L.
    Chikoidze, G.
    Russian word prediction with morphological support2004Ingår i: 5th International Symposium on Language, Logic and Computation, Tbilisi, Georgia, 2004Konferensbidrag (Refereegranskat)
    Abstract [en]

    A co-operative project between two research groups in Tbilisi and Stockholm began in1996. Its purpose is to extend a word predictor developed by the Swedish partner to the Russianlanguage. Since Russian is much richer in morphological forms than the 7 languages previouslyworked with, an additional morphological component, using an algorithm supplied by the groupin Tbilisi, is seen as necessary. It will provide inflectional categories and resulting inflectionsfor verbs, nouns and adjectives. The correct word forms can then be presented to the user of theword prediction system in a consistent manner, allowing the user to easily choose the desiredinflectional word form. At present, the work with the classification of verbs is complete. Thealgorithm is also being used to automatically tag the large lexicon used in the word predictorwith inflectional classes.

  • 348.
    Hunnicutt, Sharon
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Zuurman, M.
    Comparison of vocabulary in several symbol sets and Voice of America word list2004Ingår i: 11th Biennial Conference of the International Society for Augmentative and Alternative Communication (extended abstract), Natal, Brazil, 2004Konferensbidrag (Refereegranskat)
  • 349. Husby, O.
    et al.
    Øvregaard, Å.
    Wik, Preben
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Bech, Ø.
    Albertsen, E.
    Nefzaoui, S.
    Skarpnes, E.
    Koreman, J.
    Dealing with L1 background and L2 dialects in Norwegian CAPT2011Ingår i: Proceedings of ISCA International Workshop on Speech and Language Technology in Education, 2011Konferensbidrag (Refereegranskat)
    Abstract [en]

    This article describes the CALST project, in which the primary aim is to develop Ville-N, a CAPT system for learners of Norwegian as a second language. Since there is no accepted pronunciation standard in Norwegian, the system uses four dialects (represented by one male and one female speaker each). Ville-N makes use of L1-L2map, a tool for multi-lingual contrastive analysis, to generate a list of expected pronunciation problems. These can be used to tailor pronunciation and listening exercises. The tool can also be used for other target languages. We propose L1-L2map as a collaborative tool for the CAPT community. Index Terms. CAPT, Ville-N, Norwegian, dialects, multi-lingual contrastive analysis, L1-L2map

  • 350. Iatropoulos, G.
    et al.
    Herman, Pawel
    KTH, Skolan för elektroteknik och datavetenskap (EECS), Beräkningsvetenskap och beräkningsteknik (CST).
    Lansner, Anders
    KTH, Skolan för elektroteknik och datavetenskap (EECS), Beräkningsvetenskap och beräkningsteknik (CST).
    Karlgren, Jussi
    KTH, Skolan för elektroteknik och datavetenskap (EECS), Teoretisk datalogi, TCS. Gavagai, Slussplan 9, Stockholm, Sweden.
    Larsson, M.
    Olofsson, J. K.
    The language of smell: Connecting linguistic and psychophysical properties of odor descriptors2018Ingår i: Cognition, ISSN 0010-0277, E-ISSN 1873-7838, Vol. 178, s. 37-49Artikel i tidskrift (Refereegranskat)
    Abstract [en]

    The olfactory sense is a particularly challenging domain for cognitive science investigations of perception, memory, and language. Although many studies show that odors often are difficult to describe verbally, little is known about the associations between olfactory percepts and the words that describe them. Quantitative models of how odor experiences are described in natural language are therefore needed to understand how odors are perceived and communicated. In this study, we develop a computational method to characterize the olfaction-related semantic content of words in a large text corpus of internet sites in English. We introduce two new metrics: olfactory association index (OAI, how strongly a word is associated with olfaction) and olfactory specificity index (OSI, how specific a word is in its description of odors). We validate the OAI and OSI metrics using psychophysical datasets by showing that terms with high OAI have high ratings of perceived olfactory association and are used to describe highly familiar odors. In contrast, terms with high OSI have high inter-individual consistency in how they are applied to odors. Finally, we analyze Dravnieks's (1985) dataset of odor ratings in terms of OAI and OSI. This analysis reveals that terms that are used broadly (applied often but with moderate ratings) tend to be olfaction-unrelated and abstract (e.g., “heavy” or “light”; low OAI and low OSI) while descriptors that are used selectively (applied seldom but with high ratings) tend to be olfaction-related (e.g., “vanilla” or “licorice”; high OAI). Thus, OAI and OSI provide behaviorally meaningful information about olfactory language. These statistical tools are useful for future studies of olfactory perception and cognition, and might help integrate research on odor perception, neuroimaging, and corpus-based linguistic models of semantic organization.

45678910 301 - 350 av 692
RefereraExporteraLänk till träfflistan
Permanent länk
Referera
Referensformat
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Annat format
Fler format
Språk
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Annat språk
Fler språk
Utmatningsformat
  • html
  • text
  • asciidoc
  • rtf