Ändra sökning
Avgränsa sökresultatet
2345678 201 - 250 av 1064
RefereraExporteraLänk till träfflistan
Permanent länk
Referera
Referensformat
  • apa
  • harvard1
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Annat format
Fler format
Språk
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Annat språk
Fler språk
Utmatningsformat
  • html
  • text
  • asciidoc
  • rtf
Träffar per sida
  • 5
  • 10
  • 20
  • 50
  • 100
  • 250
Sortering
  • Standard (Relevans)
  • Författare A-Ö
  • Författare Ö-A
  • Titel A-Ö
  • Titel Ö-A
  • Publikationstyp A-Ö
  • Publikationstyp Ö-A
  • Äldst först
  • Nyast först
  • Skapad (Äldst först)
  • Skapad (Nyast först)
  • Senast uppdaterad (Äldst först)
  • Senast uppdaterad (Nyast först)
  • Disputationsdatum (tidigaste först)
  • Disputationsdatum (senaste först)
  • Standard (Relevans)
  • Författare A-Ö
  • Författare Ö-A
  • Titel A-Ö
  • Titel Ö-A
  • Publikationstyp A-Ö
  • Publikationstyp Ö-A
  • Äldst först
  • Nyast först
  • Skapad (Äldst först)
  • Skapad (Nyast först)
  • Senast uppdaterad (Äldst först)
  • Senast uppdaterad (Nyast först)
  • Disputationsdatum (tidigaste först)
  • Disputationsdatum (senaste först)
Markera
Maxantalet träffar du kan exportera från sökgränssnittet är 250. Vid större uttag använd dig av utsökningar.
  • 201.
    Bresin, Roberto
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Musikakustik.
    Delle Monache, Stefano
    University of Verona.
    Fontana, Federico
    University of Verona.
    Papetti, Stefano
    University of Verona.
    Polotti, Pietro
    University of Verona.
    Visell, Yon
    McGill University.
    Auditory feedback through continuous control of crumpling sound synthesis2008Ingår i: Proceedings of Sonic Interaction Design: Sound, Information and Experience. A CHI 2008 Workshop organized by COST Action IC0601, IUAV University of Venice , 2008, s. 23-28Konferensbidrag (Refereegranskat)
    Abstract [en]

    A realtime model for the synthesis of crumpling sounds ispresented. By capturing the statistics of short sonic transients which give rise to crackling noise, it allows for a consistent description of a broad spectrum of audible physical processes which emerge in several everyday interaction contexts.The model drives a nonlinear impactor that sonifies every transient, and it can be parameterized depending on the physical attributes of the crumpling material. Three different scenarios are described, respectively simulating the foot interaction with aggregate ground materials, augmenting a dining scenario, and affecting the emotional content of a footstep sequence. Taken altogether, they emphasize the potential generalizability of the model to situations in which a precise control of auditory feedback can significantly increase the enactivity and ecological validity of an interface.

  • 202.
    Bresin, Roberto
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Medieteknik och interaktionsdesign, MID.
    Friberg, Anders
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Musikakustik.
    Emotion rendering in music: Range and characteristic values of seven musical variables2011Ingår i: Cortex, ISSN 0010-9452, E-ISSN 1973-8102, Vol. 47, nr 9, s. 1068-1081Artikel i tidskrift (Refereegranskat)
    Abstract [en]

    Many studies on the synthesis of emotional expression in music performance have focused on the effect of individual performance variables on perceived emotional quality by making a systematical variation of variables. However, most of the studies have used a predetermined small number of levels for each variable, and the selection of these levels has often been done arbitrarily. The main aim of this research work is to improve upon existing methodologies by taking a synthesis approach. In a production experiment, 20 performers were asked to manipulate values of 7 musical variables simultaneously (tempo, sound level, articulation, phrasing, register, timbre, and attack speed) for communicating 5 different emotional expressions (neutral, happy, scary, peaceful, sad) for each of 4 scores. The scores were compositions communicating four different emotions (happiness, sadness, fear, calmness). Emotional expressions and music scores were presented in combination and in random order for each performer for a total of 5 x 4 stimuli. The experiment allowed for a systematic investigation of the interaction between emotion of each score and intended expressed emotions by performers. A two-way analysis of variance (ANOVA), repeated measures, with factors emotion and score was conducted on the participants' values separately for each of the seven musical factors. There are two main results. The first one is that musical variables were manipulated in the same direction as reported in previous research on emotional expressive music performance. The second one is the identification for each of the five emotions the mean values and ranges of the five musical variables tempo, sound level, articulation, register, and instrument. These values resulted to be independent from the particular score and its emotion. The results presented in this study therefore allow for both the design and control of emotionally expressive computerized musical stimuli that are more ecologically valid than stimuli without performance variations.

  • 203.
    Bresin, Roberto
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Musikakustik.
    Friberg, Anders
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Musikakustik.
    Evaluation of computer systems for expressive music performance2013Ingår i: Guide to Computing for Expressive Music Performance / [ed] Kirke, Alexis; Miranda, Eduardo R., Springer, 2013, s. 181-203Kapitel i bok, del av antologi (Refereegranskat)
    Abstract [en]

    In this chapter, we review and summarize different methods for the evaluation of CSEMPs. The main categories of evaluation methods are (1) comparisons with measurements from real performances, (2) listening experiments, and (3) production experiments. Listening experiments can be of different types. For example, in some experiments, subjects may be asked to rate a particular expressive characteristic (such as the emotion conveyed or the overall expression) or to rate the effect of a particular acoustic cue. In production experiments, subjects actively manipulate system parameters to achieve a target performance. Measures for estimating the difference between performances are discussed in relation to the objectives of the model and the objectives of the evaluation. There will be also a section with a presentation and discussion of the Rencon (Performance Rendering Contest). Rencon is a contest for comparing the expressive musical performances of the same score generated by different CSEMPs. Practical examples from previous works are presented, commented on, and analysed.

  • 204.
    Bresin, Roberto
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Musikakustik.
    Friberg, Anders
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Musikakustik.
    Influence of Acoustic Cues on the Expressive Performance of Music2008Ingår i: Proceedings of the 10th International Conference on Music Perception and Cognition, Sapporo, Japan, 2008Konferensbidrag (Refereegranskat)
  • 205.
    Bresin, Roberto
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Medieteknik och interaktionsdesign, MID.
    Hansen, Kjetil Falkenberg
    KTH, Skolan för datavetenskap och kommunikation (CSC), Medieteknik och interaktionsdesign, MID. KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Musikakustik.
    Dahl, Sofia
    The Radio Baton as configurable musical instrument and controller2003Ingår i: Proc. Stockholm Music Acoustics Conference, 2003, Vol. 2, s. 689-691Konferensbidrag (Refereegranskat)
    Abstract [en]

    The Max Mathews radio baton (RB) has been produced in about 40 pieces until today. It has usually been applied as an orchestra conducting system, as interactive music composition controller using typical percussionist gestures, and as a controller for sound synthesis models. In the framework of the Sounding Object EU founded project, the RB has found new applications scenarios. Three applications were based on this controller. This was achieved by changing the gesture controls. Instead of the default batons, a new radio sender that fits the fingertips was developed. This new radio sender allows musicians’ interaction based on hand gestures and it can also fit different devices. A Pd model of DJ scratching techniques (submitted to SMAC03) was controlled with the RB and the fingertip radio sender. This controller allows DJs a direct control of sampled sounds maintaining hand gestures similar to those used on vinyl. The sound model of a bodhran (submitted to SMAC03) was controlled with a traditional playing approach. The RB was controlled with a traditional bodhran double beater with one fingertip radio sender at each end. This allowed detection of the beater position on the RB surface, the surfaced corresponding to the membrane in the sound model. In a third application the fingertip controller was used to move a virtual ball rolling along the elastic surface of a box placed over the surface of the RB. The DJ console and the virtual bodhran were played in concerts.

  • 206.
    Bresin, Roberto
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Musikakustik.
    Hansen, Kjetil Falkenberg
    KTH, Skolan för datavetenskap och kommunikation (CSC), Medieteknik och interaktionsdesign, MID. KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Musikakustik.
    Karjalainen, Matti
    Helsinki University of Technology.
    Mäki-Patola, Teemu
    Helsinki University of Technology.
    Kanerva, Aki
    Helsinki University of Technology.
    Huovilainen, Antti
    Helsinki University of Technology.
    Jordá, Sergi
    University Pompeu Fabra.
    Kaltenbrunner, Martin
    University Pompeu Fabra.
    Geiger, Günter
    University Pompeu Fabra.
    Bencina, Ross
    University Pompeu Fabra.
    de Götzen, Amalia
    University of Padua.
    Rocchesso, Davide
    IUAV University of Venice.
    Controlling sound production2008Ingår i: Sound to Sense, Sense to Sound: A state of the art in Sound and Music Computing / [ed] Polotti, Pietro; Rocchesso, Davide, Berlin: Logos Verlag , 2008, s. 447-486Kapitel i bok, del av antologi (Refereegranskat)
  • 207.
    Bresin, Roberto
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH.
    Hermann, T.
    Hunt, A.
    Interactive sonification2012Ingår i: Journal on Multimodal User Interfaces, ISSN 1783-7677, E-ISSN 1783-8738, Vol. 5, nr 3-4, s. 85-86Artikel i tidskrift (Refereegranskat)
    Abstract [en]

    In October 2010, Roberto Bresin, Thomas Hermann and Andy Hunt launched a call for papers for a special issue on Interactive Sonification of the Journal on Multimodal User Interfaces (JMUI). The call was published in eight major mailing lists in the field of Sound and Music Computing and on related websites. Twenty manuscripts were submit- ted for review, and eleven of them have been accepted for publication after further improvements. Three of the papers are further developments of works presented at ISon 2010— Interactive Sonification workshop. Most of the papers went through a three-stage review process.

    The papers give an interesting overview of the field of Interactive Sonification as it is today. Their topics include the sonification of data exploration and of motion, a new sound synthesis model suitable for interactive sonification applications, a study on perception in the everyday periphery of attention, and the proposal of a conceptual framework for interactive sonification. 

  • 208.
    Bresin, Roberto
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Musikakustik.
    Hermann, ThomasBielefeld University, Bielefeld, Germany.Hunt, AndyUniversity of York, York, UK.
    Proceedings of ISon 2010 - Interactive Sonification Workshop: Human Interaction with Auditory Displays2010Proceedings (redaktörskap) (Refereegranskat)
    Abstract [en]

    Introduction

    These are the proceedings of the ISon 2010 meeting, which is the 3rd international Interactive Sonification Workshop. The first ISon workshop was held in Bielefeld (Germany) in 2004, and a second one was held in York (UK) in 2007.These meetings:

    • focus on the link between auditory displays and human‐computer interaction
    • bring together experts in sonification to exchange ideas and work‐in‐progress
    • strengthen networking in sonification research

    High quality work is assured by a peer‐reviewing process, and the successful papers were presented at the conference and are published here.

    ISon 2010 was supported by COST IC0601 Action on Sonic Interaction Design (SID) (http://www.cost‐sid.org/).

     

    About Interactive Sonification

    Sonification & Auditory Displays are increasingly becoming an established technology for exploring data, monitoring complex processes, or assisting exploration and navigation of data spaces. Sonification addresses the auditory sense by transforming data into sound, allowing the human user to get valuable information from data by using their natural listening skills.

    The main differences of sound displays over visual displays are that sound can:

    • Represent frequency responses in an instant (as timbral characteristics)
    • Represent changes over time, naturally
    • Allow microstructure to be perceived
    • Rapidly portray large amounts of data
    • Alert listener to events outside the current visual focus
    • Holistically bring together many channels of information

    Auditory displays typically evolve over time since sound is inherently a temporal phenomenon. Interaction thus becomes an integral part of the process in order to select, manipulate, excite or control the display, and this has implications for the interface between humans and computers. In recent years it has become clear that there is an important need for research to address the interaction with auditory displays more explicitly. Interactive Sonification is the specialized research topic concerned with the use of sound to portray data, but where there is a human being at the heart of an interactive control loop. Specifically it deals with:

    • interfaces between humans and auditory displays
    • mapping strategies and models for creating coherency between action and reaction (e.g. acoustic feedback, but also combined with haptic or visual feedback)
    • perceptual aspects of the display (how to relate actions and sound, e.g. cross‐modal effects, importance of synchronisation)
    • applications of Interactive Sonification
    • evaluation of performance, usability and multi‐modal interactive systems including auditory feedback

    Although ISon shines a spotlight on the particular situations where there is real‐time interaction with sonification systems, the usual community for exploring all aspects of auditory display is ICAD (http://www.icad.org/).

     

    Contents

    These proceedings contain the conference versions of all contributions to the 3rd International interactive Sonification Workshop. Where papers have audio or audiovisual examples, these are listed in the paper and will help to illustrate the multimedia content more clearly.

    We very much hope that the proceedings provide an inspiration for your work and extend your perspective on the new emerging research field of interactive sonification.

    Roberto Bresin, Thomas Hermann, Andy Hunt, ISon 2010 Organisers

  • 209. Bruce, G.
    et al.
    Schötz, S.
    Granström, Björn
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH.
    Enflo, Laura
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH.
    Modelling intonation in varieties of swedish2008Ingår i: Proceedings of the 4th International Conference on Speech Prosody, SP 2008, International Speech Communications Association , 2008, s. 571-574Konferensbidrag (Refereegranskat)
    Abstract [en]

    The research project Simulating intonational varieties of Swedish (SIMULEKT) aims to gain more precise and thorough knowledge about some major regional varieties of Swedish: South, Göta, Svea, Gotland, Dala, North, and Finland Swedish. In this research effort, the Swedish prosody model and different forms of speech synthesis play a prominent role. The two speech databases SweDia 2000 and SpeechDat constitute our main material for analysis. As a first test case for our prosody model, we compared Svea and North Swedish intonation in a pilot production-oriented perception test. Näi{dotless}ve Swedish listeners were asked to identify the most Svea and North sounding stimuli. Results showed that listeners can differentiate between the two varieties from intonation only. They also provided information on how intonational parameters affect listeners' impression of Swedish varieties. All this indicates that our experimental method can be used to test perception of different regional varieties of Swedish.

  • 210.
    Bruce, Gösta
    et al.
    Lund University.
    Schötz, Susanne
    Lund University.
    Granström, Björn
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    SIMULEKT: modelling Swedish regional intonation2007Ingår i: Proceedings of Fonetik 2007, Stockholm: KTH Royal Institute of Technology, 2007, Vol. 50, nr 1, s. 121-124Konferensbidrag (Övrigt vetenskapligt)
    Abstract [en]

    This paper introduces a new research project Simulating Intonational Varieties of Swedish (SIMULEKT). The basic goal of the project is to produce more precise and thorough knowledge about some major intonational varieties of Swedish. In this research effort the Swedish prosody model plays a prominent role. A fundamental idea is to take advantage of speech synthesis in different forms. In our analysis and synthesis work we will focus on some major intonational types: South, Göta, Svea, Gotland, Dala, North, and Finland Swedish. The significance of our project work will be within basic research as well as in speech technology applications.

  • 211.
    Brunsberg, Sandra
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Språk och kommunikation.
    Shaw, Philip
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Språk och kommunikation.
    The mathematical English of Swedish undergraduates: assimilation and adaptation2005Ingår i: Språk på tvärs: Rapport från ASLA:s höstsymposium Södertörn, 11–12 november 2004 / [ed] Boel De Geer, Anna Malmbjer, Uppsala: Svenska föreningen för tillämpad språkvetenskap, ASLA , 2005, s. 119-130.Konferensbidrag (Refereegranskat)
  • 212. Brusk, J.
    et al.
    Lager, T.
    Hjalmarsson, Anna
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Wik, Preben
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    DEAL – Dialogue Management in SCXML for Believable Game Characters2007Ingår i: Proceedings of the 2007 Conference on Future Play, Future Play '07, 2007, s. 137-144Konferensbidrag (Refereegranskat)
    Abstract [en]

    In order for game characters to be believable, they must appear to possess qualities such as emotions, the ability to learn and adapt as well as being able to communicate in natural language. With this paper we aim to contribute to the development of believable non-player characters (NPCs) in games, by presenting a method for managing NPC dialogues. We have selected the trade scenario as an example setting since it offers a well-known and limited domain common in games that support ownership, such as role-playing games. We have developed a dialogue manager in State Chart XML, a newly introduced W3C standard, as part of DEAL -- a research platform for exploring the challenges and potential benefits of combining elements from computer games, dialogue systems and language learning.

  • 213.
    Burger, Birgitta
    et al.
    Finnish Centre of Excellence in Interdisciplinary Music Research, Department of Music, University of Jyväskylä.
    Bresin, Roberto
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Musikakustik.
    Communication of Musical Expression by Means of Mobile Robot Gestures2010Ingår i: Journal on Multimodal User Interfaces, ISSN 1783-7677, E-ISSN 1783-8738, Vol. 3, nr 1, s. 109-118Artikel i tidskrift (Refereegranskat)
    Abstract [en]

    We developed a robotic system that can behave in an emotional way. A 3-wheeled simple robot with limited degrees of freedom was designed. Our goal was to make the robot displaying emotions in music performance by performing expressive movements. These movements have been compiled and programmed based on literature about emotion in music, musicians’ movements in expressive performances, and object shapes that convey different emotional intentions. The emotions happiness, anger, and sadness have been implemented in this way. General results from behavioral experiments show that emotional intentions can be synthesized, displayed and communicated by an artificial creature, also in constrained circumstances.

  • 214.
    Bälter, Olle
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Människa-datorinteraktion, MDI.
    Engwall, Olov
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH. KTH, Skolan för datavetenskap och kommunikation (CSC), Centra, Centrum för Talteknologi, CTT.
    Öster, Anne-Marie
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH. KTH, Skolan för datavetenskap och kommunikation (CSC), Centra, Centrum för Talteknologi, CTT.
    Kjellström, Hedvig
    KTH, Skolan för datavetenskap och kommunikation (CSC), Datorseende och robotik, CVAP.
    Wizard-of-Oz Test of ARTUR - a Computer-Based Speech Training System with Articulation Correction2005Ingår i: proceedings of ASSETS 2005, 2005, s. 36-43Konferensbidrag (Refereegranskat)
    Abstract [en]

    This study has been performed in order to test the human-machine interface of a computer-based speech training aid named ARTUR with the main feature that it can give suggestions on how to improve articulation. Two user groups were involved: three children aged 9-14 with extensive experience of speech training, and three children aged 6. All children had general language disorders. The study indicates that the present interface is usable without prior training or instructions, even for the younger children, although it needs some improvement to fit illiterate children. The granularity of the mesh that classifies mispronunciations was satisfactory, but can be developed further.

  • 215. Camargo, Zuleica
    et al.
    Salomão, Gláucia Laís
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH.
    Pinho, Sílvia
    Análise acústica e aerodinâmica da voz (Acoustic and aerodynamic assessment of voice)2011Ingår i: Tratado de Otorrinolaringologia (Treatise on Otorhinolaryngology) / [ed] Caldas, S.; Melo, J.F.; Martins, R.H.; Selaimen, S., Sao Paulo: Editora Roca , 2011, s. 794-804Kapitel i bok, del av antologi (Övrigt vetenskapligt)
  • 216. Camurri, A.
    et al.
    Bevilacqua, F.
    Bresin, Roberto
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Musikakustik.
    Maestre, E.
    Penttinen, H.
    Seppänen, J.
    Välimäki, V.
    Volpe, G.
    Warusfel, O.
    Embodied music listening and making in context-aware mobile applications: the EU-ICT SAME Project2009Konferensbidrag (Refereegranskat)
  • 217. Camurri, A.
    et al.
    De Poli, G.
    Friberg, Anders
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Musikakustik.
    Leman, M.
    Volpe, G.
    The MEGA project: Analysis and synthesis of multisensory expressive gesture in performing art applications2005Ingår i: Journal of New Music Research, ISSN 0929-8215, E-ISSN 1744-5027, Vol. 34, nr 1, s. 5-21Artikel i tidskrift (Refereegranskat)
    Abstract [en]

    This article presents a survey of the research work carried out within the framework of the European Union-IST project MEGA (Multisensory Expressive Gesture Applications, November 2000-October 2003; www. megaproject.org). First, the article introduces a layered conceptual framework for analysis and synthesis of expressive gesture. Such a framework represents the main methodological foundation upon which the MEGA project built its own research. A brief overview of the achievements of research in expressive gesture analysis and synthesis is then provided: these are the outcomes of some experiments that were carried out in order to investigate specific aspects of expressive gestural communication. The work resulted in the design and development of a collection of software libraries integrated in the MEGA System Environment (MEGASE) based on the EyesWeb open platform (www. eyesweb.org).

  • 218.
    Camurri, Antonio
    et al.
    University of Genova.
    Volpe, Gualtiero
    University of Genova.
    Vinet, Hugues
    IRCAM, Paris.
    Bresin, Roberto
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Musikakustik.
    Fabiani, Marco
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH.
    Dubus, Gaël
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Musikakustik.
    Maestre, Esteban
    Universitat Pompeu Fabra, Barcelona.
    Llop, Jordi
    Universitat Pompeu Fabra, Barcelona.
    Kleimola, Jari
    Oksanen, Sami
    Välimäki, Vesa
    Seppanen, Jarno
    User-centric context-aware mobile applications for embodied music listening2009Ingår i: User Centric Media / [ed] Akan, Ozgur; Bellavista, Paolo; Cao, Jiannong; Dressler, Falko; Ferrari, Domenico; Gerla, Mario; Kobayashi, Hisashi; Palazzo, Sergio; Sahni, Sartaj; Shen, Xuemin (Sherman); Stan, Mircea; Xiaohua, Jia; Zomaya, Albert; Coulson, Geoffrey; Daras, Petros; Ibarra, Oscar Mayora, Heidelberg: Springer Berlin , 2009, s. 21-30Kapitel i bok, del av antologi (Refereegranskat)
    Abstract [en]

    This paper surveys a collection of sample applications for networked user-centric context-aware embodied music listening. The applications have been designed and developed in the framework of the EU-ICT Project SAME (www.sameproject.eu) and have been presented at Agora Festival (IRCAM, Paris, France) in June 2009. All of them address in different ways the concept of embodied, active listening to music, i.e., enabling listeners to interactively operate in real-time on the music content by means of their movements and gestures as captured by mobile devices. In the occasion of the Agora Festival the applications have also been evaluated by both expert and non-expert users

  • 219.
    Carlson, Rolf
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Conflicting acoustic cues in stop perception2007Ingår i: Where Do Features Come From ?: Phonological Primitives in the Brain, the Mouth, and the Ear, 2007, s. 63-64Konferensbidrag (Refereegranskat)
  • 220.
    Carlson, Rolf
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Using acoustic cues in stop perception2007Ingår i: Proceedings of Fonetik 2007, 2007, Vol. 50, nr 1, s. 25-28Konferensbidrag (Övrigt vetenskapligt)
  • 221.
    Carlson, Rolf
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Edlund, Jens
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Heldner, Mattias
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Hjalmarsson, Anna
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    House, David
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Skantze, Gabriel
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Towards human-like behaviour in spoken dialog systems2006Ingår i: Proceedings of Swedish Language Technology Conference (SLTC 2006), Gothenburg, Sweden, 2006Konferensbidrag (Övrigt vetenskapligt)
    Abstract [en]

    We and others have found it fruitful to assume that users, when interacting with spoken dialogue systems, perceive the systems and their actions metaphorically. Common metaphors include the human metaphor and the interface metaphor (cf. Edlund, Heldner, & Gustafson, 2006). In the interface metaphor, the spoken dialogue system is perceived as a machine interface – often but not always a computer interface. Speech is used to accomplish what would have otherwise been accomplished by some other means of input, such as a keyboard or a mouse. In the human metaphor, on the other hand, the computer is perceived as a creature (or even a person) with humanlike conversational abilities, and speech is not a substitute or one of many alternatives, but rather the primary means of communicating with this creature. We are aware that more “natural ” or human-like behaviour does not automatically make a spoken dialogue system “better ” (i.e. more efficient or more well-liked by its users). Indeed, we are quite convinced that the advantage (or disadvantage) of humanlike behaviour will be highly dependent on the application. However, a dialogue system that is coherent with a human metaphor may profit from a number of characteristics.

  • 222.
    Carlson, Rolf
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Elenius, Kjell
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Swerts, Marc
    Tilburg University, The Netherlands.
    Perceptual judgments of pitch range2004Ingår i: Proc. of Intl Conference on Speech Prosody 2004 / [ed] Bel, B.; Marlin, I., Nara, Japan, 2004, s. 689-692Konferensbidrag (Refereegranskat)
    Abstract [en]

    This paper reports on a study that explores to what extent listeners are able to judge where a particular utterance fragment is located in a speaker's pitch range. The research consists of a perception study that makes use of 100 stimuli, selected from 50 different speakers whose speech was originally collected for a multi-speaker database of Swedish speech materials. The fragments are presented to subjects whom are asked to estimate whether the fragment is located in the lower or higher part of that speaker's range. Results reveal that listeners' judgments are dependent on the gender of the speaker, but that within a gender they tend to hear differences in range.

  • 223.
    Carlson, Rolf
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Granstrom, Bjorn
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Johan Liljencrants (1936-2012) in memoriam2012Ingår i: Journal of the International Phonetic Association, ISSN 0025-1003, E-ISSN 1475-3502, Vol. 42, nr 2, s. 253-254Artikel i tidskrift (Refereegranskat)
  • 224.
    Carlson, Rolf
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Granström, Björn
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Data-driven multimodal synthesis2005Ingår i: Speech Communication, ISSN 0167-6393, E-ISSN 1872-7182, Vol. 47, nr 02-jan, s. 182-193Artikel i tidskrift (Refereegranskat)
    Abstract [en]

    This paper is a report on current efforts at the Department of Speech, Music and Hearing, KTH, on data-driven multimodal synthesis including both visual speech synthesis and acoustic modeling. In the research we try to combine both corpus based methods with knowledge based models and to explore the best of the two approaches. In the paper an attempt to build formant-synthesis systems based on both rule-generated and database driven methods is presented. A pilot experiment is also reported showing that this approach can be a very interesting path to explore further. Two studies on visual speech synthesis are reported, one on data acquisition using a combination of motion capture techniques and one concerned with coarticulation, comparing different models.

  • 225.
    Carlson, Rolf
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Granström, Björn
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Rule-based Speech Synthesis2008Ingår i: Springer Handbook of Speech Processing / [ed] Benesty, J.; Sondhi, M. M.; Huang, Y., Berlin/Heidelberg: Springer Berlin/Heidelberg, 2008, s. 429-436Kapitel i bok, del av antologi (Refereegranskat)
    Abstract [en]

    In this chapter, we review some of the issues in rule-based synthesis and specifically discuss formant synthesis. Formant synthesis and the theory behind have played an important role in both the scientific progress in understanding how humans talk and also the development of the first speech technology applications. Its flexibility and small footprint makes the approach still of interest and a valuable complement to the current dominant methods based on concatenative data-driven synthesis. As already mentioned in the overview by Schroeter (Chap. 19) we also see a new trend to combine the rule-based and data-driven approaches. Formant features from a database that can be used both to optimize a rule-based formant synthesis system and to optimize the search for good units in a concatenative system.

  • 226.
    Carlson, Rolf
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH.
    Granström, Björn
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH.
    Speech Synthesis2010Ingår i: The Handbook of Phonetic Sciences, Blackwell Publishing, 2010, 2, s. 781-803Kapitel i bok, del av antologi (Refereegranskat)
  • 227.
    Carlson, Rolf
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH.
    Granström, Björn
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH.
    Lindblom, Björn
    Risberg, Arne
    Sundberg, Johan
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH.
    Gunnar Fant 1920-2009 In Memoriam2009Ingår i: Phonetica, ISSN 0031-8388, E-ISSN 1423-0321, Vol. 66, nr 4, s. 249-250Artikel i tidskrift (Refereegranskat)
  • 228.
    Carlson, Rolf
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Gustafson, Kjell
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Exploring Data Driven Parametric Synthesis2009Ingår i: Proceedings of Fonetik 2009: The XXIIth Swedish Phonetics Conference / [ed] Peter Branderud, Hartmut Traunmüller, Stockholm, Sweden: Stockholm University, 2009, s. 86-91Konferensbidrag (Övrigt vetenskapligt)
    Abstract [en]

    This paper describes our work on building aformant synthesis system based on both rule generated and database driven methods. Three parametric synthesis systems are discussed: our traditional rule based system, a speaker adapted system, and finally a gesture system.The gesture system is a further development of the adapted system in that it includes concatenated formant gestures from a data-driven unit library. The systems are evaluated technically, comparing the formant tracks with an analysed test corpus. The gesture system results in a 25% error reduction in the formant frequencies due to the inclusion of the stored gestures. Finally, a perceptual evaluation shows a clear advantage in naturalness for the gesture system compared to both the traditional system and the speaker adapted system.

  • 229.
    Carlson, Rolf
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Gustafson, Kjell
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Strangert, E.
    Synthesising disfluencies in a dialogue system2009Ingår i: Nordic Prosody: Proccedings of the Xth Conference / [ed] Vainio, M., Aulanko, R., Aaltonen, O., Frankfurt am Main: Peter Lang Publishing Group, 2009Konferensbidrag (Refereegranskat)
  • 230.
    Carlson, Rolf
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Gustafson, Kjell
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Strangert, Eva
    Cues for Hesitation in Speech Synthesis2006Ingår i: INTERSPEECH 2006 AND 9TH INTERNATIONAL CONFERENCE ON SPOKEN LANGUAGE PROCESSING, BAIXAS: ISCA-INST SPEECH COMMUNICATION ASSOC , 2006, s. 1300-1303Konferensbidrag (Refereegranskat)
    Abstract [en]

    The current study investigates acoustic correlates to perceived hesitation based on previous work showing that pause duration and final lengthening both contribute to the perception of hesitation. It is the total duration increase that is the valid cue rather than the contribution by either factor. The present experiment using speech synthesis was designed to evaluate F0 slope and presence vs. absence of creaky voice before the inserted hesitation in addition to durational cues. The manipulations occurred in two syntactic positions, within a phrase and between two phrases, respectively. The results showed that in addition to durational increase, variation of both F0 slope and creaky voice had perceptual effects, although to a much lesser degree. The results have a bearing on efforts to model spontaneous speech including disfluencies, to be explored, for example, in spoken dialogue systems.

  • 231.
    Carlson, Rolf
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Gustafson, Kjell
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Strangert, Eva
    Umeå University.
    Prosodic Cues for Hesitation2006Ingår i: Working Papers 52: Proceedings from Fonetik 2006 / [ed] Gilbert Ambrazaitis, Susanne Schötz, Lund: Lund University , 2006, Vol. 52, s. 21-24Konferensbidrag (Övrigt vetenskapligt)
    Abstract [en]

    In our efforts to model spontaneous speech for use in, for example, spoken dialogue systems, a series of experiments have been conducted in order to investigate correlates to perceived hesitation. Previous work has shown that it is the total duration increase that is the valid cuerather than the contribution by either of the two factors pause duration and final lengthening. In the present experiment we explored the effects of F0 slope variation and the presence vs. absence of creaky voice in addition to durational cues, using synthetic stimuli. The results showed that variation of both F0 slope and creaky voice did have perceptual effects, but to amuch lesser degree than the durational increase.

  • 232.
    Carlson, Rolf
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Gustafsson, Kjell
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Strangert, Eva
    Umeå University.
    Modelling hesitation for synthesis of spontaneous speech2006Ingår i: Proceedings of Speech Prosody 2006 / [ed] R. Hoffmann, H. Mixdorff, Dresden, 2006Konferensbidrag (Refereegranskat)
    Abstract [en]

    The current work deals with the modelling of one type of disfluency, hesitations. A perceptual experiment using speech synthesis was designed to evaluate two duration features found to be correlates to hesitation, pause duration and final lengthening. A variation of F0 slope before the hesitation wasalso included. The most important finding is that it is the totalduration increase that is the valid cue rather than the contribution by either factor. In addition, our findings lead us to assume an interaction with syntax. The absence of strong effects of the induced F0 variation was unexpected and we consider several possible explanations for this result.

  • 233.
    Carlson, Rolf
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Hawkins, Sarah
    University of Cambridge.
    When is fine phonetic detail a detail?: 16th International Congress of Phonetics Sciences2007Ingår i: Proceedings of ICPhS 2007 / [ed] Jürgen Trouvain, William J. Barry, Saarbrücken, Germany, 2007, s. 211-214Konferensbidrag (Refereegranskat)
  • 234.
    Carlson, Rolf
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Hirschberg, J.
    Swerts, M.
    Cues to upcoming Swedish prosodic boundaries: Subjective judgment studies and acoustic correlates2005Ingår i: Speech Communication, ISSN 0167-6393, E-ISSN 1872-7182, Vol. 46, nr 3-4, s. 326-333Artikel i tidskrift (Refereegranskat)
    Abstract [en]

    Studies of perceptually based predictions of upcoming prosodic boundaries in spontaneous Swedish speech, both by native speakers of Swedish and of native speakers of standard American English reveal marked similarity in judgments. We examined whether Swedish and American listeners were able to predict the occurrence and strength of upcoming boundaries in a series of web-based perceptive experiments. Utterance fragments (in both long and short versions) were selected from a corpus of spontaneous Swedish speech, which was first labeled for boundary presence and strength by expert labelers. These fragments were then presented to listeners, who were instructed to guess whether or not they were followed by a prosodic break, and if so, what the strength of the break was. Results revealed that both Swedish and American listening groups were indeed able to predict whether or not a boundary (of a particular strength) followed the fragment. This suggests that acoustic and prosodic, rather than lexico-grammatical and semantic information was being used by listeners as a primary cue. Acoustic and prosodic correlates of these judgments were then examined, with significant correlations found between judgments and the presence/absence of final creak and phrase-final f0 level and slope.

  • 235.
    Carlson, Rolf
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Hirschberg, Julia
    Columbia University.
    Swerts, Marc
    University of Tilburg, The Netherlands.
    Prediction of upcoming Swedish prosodic boundaries by Swedish and American listeners2004Ingår i: Proc of Intl Conference on Speech Prosody 2004 / [ed] Bel, B.; Marlin, I., Nara, Japan, 2004, s. 329-332Konferensbidrag (Refereegranskat)
    Abstract [en]

    We describe results of a study of perceptually based predictions of upcoming prosodic breaks in spontaneous Swedish speech materials by native speakers of Swedish and of standard American English. The question addressed here is the extent to which listeners are able, on the basis of acoustic and prosodic features, to predict the occurrence of upcoming boundaries, and if so, whether they are able to distinguish different degrees of boundary strength. An experiment was conducted in which spontaneous utterance fragments (both long and short versions) were presented to listeners, who were instructed to guess whether or not the fragments were followed by a prosodic break, and if so, what the strength of the break was, where boundary presence and strength had been independently labeled. Results revealed that both listening groups were indeed able to predict whether or not a boundary (of a particular strength) followed the fragment, suggesting that prosodic rather than lexico-grammatical information was being used as a primary cue.

  • 236.
    Carlsson, Christoffer
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH.
    The Acoustics of Stockholm Concert Hall and Artificial Reverberation Systems: Evaluation of Stora salen and simulation of its electronic reverberation system2015Självständigt arbete på avancerad nivå (masterexamen), 20 poäng / 30 hpStudentuppsats (Examensarbete)
    Abstract [sv]

    Det här examensarbetet undersöker påverkan på de akustiska egenskaperna hos en konsertlokal orsakad av ett artificiellt efterklangssystem. Likaså undersöks möjligheterna för att simulera dessa akustiska egenskaper. Genom att undersöka Stockholms konserthus, som nyligen installerade ett efterklangssystem, kommer en bättre förståelse för artificiella efterklangssystem skapas och ytterligare förbättringar för simulering kommer att möjliggöras. Den här studien genomförs i två delar: (1) objektiv data, inhämtad från akustiska mätningar, utvärderas både internt och mot andra konsertlokaler samt (2) genom datorsimulering av konsertlokalen och det elektroniska efterklangssystemet utvärderas de akustiska egenskaperna.

    Studien visar att inverkan på de akustiska egenskaperna hos Stockholms konserthus orsakade av det artificiella efterklangssystemet inte är överdrivna men noterbara. En önskad ökning av efterklangstiden med 0.3 sekunder uppnås men detta på bekostnad av att ljudets klarhet minskar med 0.7 decibel. Vidare är det möjligt att simulera ljudutbredningen i en konsertlokal som har ett efterklangssystem installerat med ett tämligen realistiskt resultat. För att uppnå detta simuleringsresultat skapas ett skript vilket väger samman alla överföringsfunktioner mellan ljudkällan och mottagaren, inklusive de mellan efterklangssystemets mikrofoner och högtalare.

  • 237. Castellana, Antonella
    et al.
    Selamtzis, Andreas
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH.
    Salvi, Giampiero
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH.
    Carullo, Alessio
    Astolfi, Arianna
    Cepstral and entropy analyses in vowels excerpted from continuous speech of dysphonic and control speakers2017Ingår i: Proceedings of the Annual Conference of the International Speech Communication Association, Interspeech 2017 / [ed] ISCA, International Speech Communication Association, 2017, Vol. 2017, s. 1814-1818Konferensbidrag (Refereegranskat)
    Abstract [en]

    There is a growing interest in Cepstral and Entropy analyses of voice samples for defining a vocal health indicator, due to their reliability in investigating both regular and irregular voice signals. The purpose of this study is to determine whether the Cepstral Peak Prominence Smoothed (CPPS) and Sample Entropy (SampEn) could differentiate dysphonic speakers from normal speakers in vowels excerpted from readings and to compare their discrimination power. Results are reported for 33 patients and 31 controls, who read a standardized phonetically balanced passage while wearing a head mounted microphone. Vowels were excerpted from recordings using Automatic Speech Recognition and, after obtaining a measure for each vowel, individual distributions and their descriptive statistics were considered for CPPS and SampEn. The Receiver Operating Curve analysis revealed that the mean of the distributions was the parameter with the highest discrimination power for both CPPS and SampEn. CPPS showed a higher diagnostic precision than SampEn, exhibiting an Area Under Curve (AUC) of 0.85 compared to 0.72. A negative correlation between the parameters was found (Spearman; p = - 0.61), with higher SampEn corresponding to lower CPPS. The automatic method used in this study could provide support to voice monitorings in clinic and during individual's daily activities.

  • 238.
    Castellano, Ginevra
    et al.
    InfoMus Lab, DIST, University of Genova.
    Bresin, Roberto
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Musikakustik.
    Camurri, Antonio
    InfoMus Lab, DIST, University of Genova.
    Volpe, Gualtiero
    InfoMus Lab, DIST, University of Genova.
    Expressive Control of Music and Visual Media by Full-Body Movement2007Ingår i: Proceedings of the 7th International Conference on New Interfaces for Musical Expression, NIME '07, New York, NY, USA: ACM Press, 2007, s. 390-391Konferensbidrag (Refereegranskat)
    Abstract [en]

    In this paper we describe a system which allows users to use their full-body for controlling in real-time the generation of an expressive audio-visual feedback. The system extracts expressive motion features from the user’s full-body movements and gestures. The values of these motion features are mapped both onto acoustic parameters for the real-time expressive rendering ofa piece of music, and onto real-time generated visual feedback projected on a screen in front of the user.

  • 239. Castellano, Ginevra
    et al.
    Bresin, Roberto
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Musikakustik.
    Camurri, Antonio
    Volpe, Gualtiero
    User-Centered Control of Audio and Visual Expressive Feedback by Full-Body Movements2007Ingår i: Affective Computing and Intelligent Interaction / [ed] Paiva, Ana; Prada, Rui; Picard, Rosalind W., Berlin / Heidelberg: Springer Berlin/Heidelberg, 2007, s. 501-510Kapitel i bok, del av antologi (Refereegranskat)
    Abstract [en]

    In this paper we describe a system allowing users to express themselves through their full-body movement and gesture and to control in real-time the generation of an audio-visual feedback. The systems analyses in real-time the user’s full-body movement and gesture, extracts expressive motion features and maps the values of the expressive motion features onto real-time control of acoustic parameters for rendering a music performance. At the same time, a visual feedback generated in real-time is projected on a screen in front of the users with their coloured silhouette, depending on the emotion their movement communicates. Human movement analysis and visual feedback generation were done with the EyesWeb software platform and the music performance rendering with pDM. Evaluation tests were done with human participants to test the usability of the interface and the effectiveness of the design.

  • 240. Caudery, T.
    et al.
    Petersen, M.
    Shaw, Philip
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Språk och kommunikation.
    The Language Environments of Exchange Students at Scandinavian Universities2007Ingår i: Researching Content and Language Integration in Higher Education / [ed] Wilkinson, R.; Zegers, V., University of Maastricht , 2007, s. 233-250Kapitel i bok, del av antologi (Refereegranskat)
    Abstract [en]

    Exchange students who come to Scandinavia are often motivated by an intention to improve their proficiency in English rather than the local language. They take academic classes conducted in English and may find themselves living in a lingua-franca English bubble, acculturated to an international-student subculture. A few do break out of the bubble, learn the local language, and experience the local culture. Here we report on a project intended identify the factors leading to successful learning of both English and the local languages. 70 students at each of four institutions, two in Sweden, two in Denmark, were interviewed three times over a semester and asked to complete simple language tests. English proficiency improved in most cases, Swedish/Danish was only learnt by those with good initial English and appropriate motivation. As expected, contact with local students was limited. Institutional policies can probably influence these outcomes.

  • 241. Caudery, T.
    et al.
    Petersen, M.
    Shaw, Philip
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Språk och kommunikation.
    The motivations of exchange students at Scandinavian universities2008Ingår i: Students, Staff and Academic Mobility in Higher Education / [ed] Byram, M.; Dervin, F., Newcastle: Cambridge Scholars Press , 2008, s. 114-130Kapitel i bok, del av antologi (Refereegranskat)
  • 242.
    Cerrato, Loredana
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    A coding scheme for the annotation of feedback phenomena in conversational speech2004Ingår i: Proc of LREC Workshop on Models of Human Behaviour for the Specification and Evaluation of Multimodal Input and Output Interfaces / [ed] Martin, J.C., Lisboa, 2004, s. 25-28Konferensbidrag (Refereegranskat)
    Abstract [en]

    A coding scheme specifically developed to label feedback phenomena in conversational speech is presented in this paper. The coding scheme allows the categorization of feedback phenomena according to their typology, direction, and communicative function in the given context. The results of the reliability tests run to verify the appropriateness of the coding scheme to code feedback phenomena in different languages and across different modalities are also presented.

  • 243.
    Cerrato, Loredana
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    A comparative study of verbal feedback in Italian and Swedish map-task dialogues2004Ingår i: Proceedings of the Nordic Symposium on the comparison of spoken languages, Copenhagen Working Papers in LSP / [ed] Copenhagen, P.; Hernrichsen, J., 2004, s. 99-126Konferensbidrag (Övrigt vetenskapligt)
  • 244.
    Cerrato, Loredana
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Linguistic functions of head nods2005Ingår i: Proceedings from The Second Nordic Conference on Multimodal Communication / [ed] Allwood, J.; Dorriots, B., Göteborg: Göteborg University , 2005, s. 137-152Konferensbidrag (Refereegranskat)
    Abstract [en]

    The aim of the present study is to investigate which communicative functions head nods can have in spoken Swedish. By nod is here meant a vertical down-up movement of the head. To classify the communicative functions of head nods 10 short video-recorded Swedish dialogues were analysed and labeled. The labels used are referred to the different communicative functions that the head nods carry out in the given context. The results show that the most common function carried out by head nods is that of feedback. Beside feedback function, head nods can be produced to signal turn taking, focus and emphasis, to give affirmative responses and to show courtesy. The visual information carried out by head nods in spoken communicative interactions is without doubt extremely important; therefore it should be exploited in the field of human-machine interfaces. This could be done by integrating head nods in the design and development of embodied conversational agents. Thanks to the production of head nods, embodied conversational agents might become more effective and appear more natural during their interactions with human beings.

  • 245.
    Cerrato, Loredana
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    On the acoustic, prosodic and gestural characteristics of “m-like” sounds in Swedish2005Ingår i: Feedback in spoken interaction: NordTalk Symposium / [ed] Jens Allwood, Göteborg: Göteborg University , 2005, Vol. Feedback in Spoken Interaction- Nordtalk Symposium 2003, s. 18-31Konferensbidrag (Refereegranskat)
    Abstract [en]

    The aim of the present study is to verify what communicative functions “m-like” sounds can have in spoken Swedish and investigate both the relationship between prosodic variation and communicative function and the relationship between the production of “mlike” sounds and their accompanying gestures. The main hypothesis tested is that the different communicative functions carried by these “m-like” sounds are conveyed by means of different prosodic cues. To test this hypothesis, audio-recordings of two dialogues, elicited with the map-task technique, were used. A distributional and functional analysis of “m-like” sounds was first carried out. Afterwards, an acoustic analysis of these sounds was performed to find out how prosodic variation and communicative function are related. The results show that the most common function carried out by “m-like” sounds is that of feedback. The general category of feedback can be further divided in sub-categories depending on the specific function that the short expression carries out in the given context. To each function it is possible to relate a prototypical F0 contour and acoustic characteristics. For the analysis of the accompanying gestures of “m-like” sounds, two AV recordings of spontaneous dialogues were used. The results of the distributional analysis show that 41% of all the analysed “m-like” sounds are accompanied by a gesture. The most common accompanying gestures are head movement s such as nods and jerks. The relationship between the function carried by speech and the specific function of the accompanying gesture has also been coded and analyzed. Gestures co-occurring with speech can either have a “non-marked/neutral” function, which means that they do not add further information to what is being said with speech, or can be produced to add, emphasize weaken or contradicting speech. When the function of these gestures is neutral, they tend to have a minimal extent, while when their specific function is to emphasize the information expressed by speech, their extent tends to be bigger. This result might be related to the fact that gestures are often produced to emphasize information that is also focused by mechanisms like prosody in speech.

  • 246.
    Cerrato, Loredana
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    The communicative function of "sì" in Italian and "ja" in Swedish: an acoustic analysis2005Ingår i: Proceedings of Fonetik 2005 / [ed] Anders Eriksson, Jonas Lindh, Göteborg: Göteborg University , 2005, s. 41-44Konferensbidrag (Övrigt vetenskapligt)
    Abstract [en]

    The results of an acoustic analysis and a perceptual evaluation of the role of prosody inspontaneously produced “ja” and “sì” in Swedish and Italian are reported and discussedin this paper. The hypothesis is that pitch contour, duration cues and relative intensity can beuseful in the identification of the different communicative functions of these short expressions taken out of their context. The results of the perceptual tests run to verify whether the acoustic cues alone can be used to distinguish different functions of the same lexical items are encouraging only for Italian “sí”, while for Swedish “ja” they show some confusions among the different categories.

  • 247.
    Cerrato, Loredana
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Ekeklint, Susanne
    Evaluating users reactions to human-like interfaces: Prosodic and paralinguistic features as new evaluation measures for users satisfaction2004Ingår i: From Brows to Trust: Evaluating Embodied Conversational Agents / [ed] Ruttkay, Z.; Pelachaud, C., Dordrecht: Kluwer Academic Publishers, 2004, s. 101-124Kapitel i bok, del av antologi (Refereegranskat)
    Abstract [en]

    An increasing number of dialogue systems are deployed to provide publicservices in our everyday lives. They are becoming more service-minded and several ofthem provide different channels for interaction. The rationale is to make automaticservices available in new environments and more attractive to use. From a developerperspective, this affects the complexity of the requirements elicitation activity, as newcombinations and variations in end-user interaction need to be considered. The aimof our investigation is to propose new parameters and metrics to evaluate multimodaldialogue systems endowed with embodied conversational agents (ECAs). These newmetrics focus on the users, rather than on the system. Our assumption is that theintentional use of prosodic variation and the production of communicative non-verbalbehaviour by users can give an indication of their attitude towards the system andmight also help to evaluate the users’ overall experience of the interaction. To testour hypothesis we carried out analyses on different Swedish corpora of interactionsbetween users and multimodal dialogue systems. We analysed the prosodic variationin the way the users ended their interactions with the system and we observed theproduction of non-verbal communicative expressions by users. Our study supports theidea that the observation of users’ prosodic variation and production of communicativenon-verbal behaviour during the interaction with dialogue systems could be used asan indication of whether or not the users are satisfied with the system performance.

  • 248.
    Cerrato, Loredana
    et al.
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Svanfeldt, Gunilla
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    A method for the detection of communicative head nods in expressive speech2006Ingår i: Papers from the Second Nordic Conference on Multimodal Communication 2005 / [ed] Allwood, J.; Dorriots, B.; Nicholson, S., Göteborg: Göteborg University , 2006, s. 153-165Konferensbidrag (Refereegranskat)
    Abstract [en]

    The aim of this study is to propose a method for automatic detection of head nods during the production of semi-spontaneous speech. This method also provides means for extracting certain characteristics of head nods, that may vary depending on placement, function and even underlying emotional expression. The material used is part of the Swedish PF-Star corpora which were recorded by means of an optical motion capture system (Qualisys) able to successfully register articulatory movements as well as head movements and facial expressions. The material consists of short sentences as well as of dialogic speech produced by a Swedish actor. The method for automatic head nods detection on the 3D data acquired with Qualisys is based on criteria for slope, amplitude and a minimum number of consecutive frames. The criteria are tuned on head nods that have been manually annotated. These parameters can be varied to detect different kinds of head movements and can also be combined with other parameters in order to detect facial gestures, such as eyebrow displacements. For this study we focused in particular on the detection of head nods, since in earlier studies they have been found to be important visual cues in particular for signaling feedback and focus. In order to evaluate the method a preliminary test was run on semi-spontaneous dialogic speech, which is also part of the Swedish PF-Star corpora and produced by the same actor who read the sentences. The results show that the parameters and the criteria that had been set on the basis of the training corpus are valid also for the dialogic speech, even if more sophisticated parameters could be useful to achieve a more precise result.

  • 249. Chollet, M.
    et al.
    Stefanov, Kalin
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH.
    Prendinger, H.
    Scherer, S.
    Public Speaking Training with a Multimodal Interactive Virtual Audience Framework2015Ingår i: ICMI '15 Proceedings of the 2015 ACM on International Conference on Multimodal Interaction, ACM Digital Library, 2015, s. 367-368Konferensbidrag (Refereegranskat)
    Abstract [en]

    We have developed an interactive virtual audience platform for public speaking training. Users' public speaking behavior is automatically analyzed using multimodal sensors, and ultimodal feedback is produced by virtual characters and generic visual widgets depending on the user's behavior. The flexibility of our system allows to compare different interaction mediums (e.g. virtual reality vs normal interaction), social situations (e.g. one-on-one meetings vs large audiences) and trained behaviors (e.g. general public speaking performance vs specific behaviors).

  • 250.
    Contardo, Ivonne
    et al.
    Karolinska Institutet, Sweden.
    McAllister, Anita
    Karolinska Institutet, Sweden.
    Strömbergsson, Sofia
    KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH, Tal-kommunikation.
    Real-time registration of listener reactions to unintelligibility in misarticulated child speech2014Ingår i: Proceedings from FONETIK 2014 / [ed] Heldner, M., Stockholm, 2014, s. 127-132Konferensbidrag (Övrigt vetenskapligt)
    Abstract [en]

    This study explores the relation between misarticulations and their impact on intelligibility. 30 listeners (17 clinicians and 13 untrained listeners) were given the task of clicking a button whenever they perceived something unintelligible during playback of misarticulated child speech samples. No differences were found between the clinicians and the untrained listeners regarding clicking frequency. The distribution of listener clicks correlated strongly with the clinical evaluations of the same samples. The distribution of clicks was also related to manually annotated speech errors, allowing examination of links between events in the speech signal and reactions evoked in listeners. Hereby, we demonstrate a viable approach to ranking speech error types with regards to their impact on intelligibility in conversational speech.

2345678 201 - 250 av 1064
RefereraExporteraLänk till träfflistan
Permanent länk
Referera
Referensformat
  • apa
  • harvard1
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Annat format
Fler format
Språk
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Annat språk
Fler språk
Utmatningsformat
  • html
  • text
  • asciidoc
  • rtf