


default search action
6th SSW 2007: Bonn, Germany
- Petra Wagner, Julia Abresch, Stefan Breuer, Wolfgang Hess:

Sixth ISCA Workshop on Speech Synthesis, SSW 2007, Bonn, Germany, August 22-24, 2007. ISCA 2007
Keynote 1
- Bernd J. Kröger:

Perspectives for articulatory speech synthesis.
Various Topics
- Oxana Govokhina, Gérard Bailly, Gaspard Breton:

Learning optimal audiovisual phasing for an HMM-based control model for facial animation. 1-4 - Peter Birkholz, Ingmar Steiner, Stefan Breuer:

Control concepts for articulatory speech synthesis. 5-10 - Alexander Kain, Qi Miao, Jan P. H. van Santen:

Spectral control in concatenative speech synthesis. 11-16 - Barry Kirkpatrick, Darragh O'Brien, Ronan Scaife:

Feature transformation applied to the detection of discontinuities in concatenated speech. 17-21
Expressive Speech Synthesis
- Nick Campbell:

Towards conversational speech synthesis; lessons learned from the expressive speech processing project. 22-27 - Shinsuke Sakai, Jinfu Ni, Ranniery Maia, Keiichi Tokuda, Minoru Tsuzaki, Tomoki Toda, Hisashi Kawai, Satoshi Nakamura:

Communicative speech synthesis with XIMERA: a first step. 28-33 - Raul Fernandez, Bhuvana Ramabhadran:

Automatic exploration of corpus-specific properties for expressive text-to-speech: a case study in emphasis. 34-39 - Charlotte Wollermann, Eva Lasarcyk:

Modeling and perceiving of (un-)certainty in articulatory speech synthesis. 40-45 - Lijuan Wang, Min Chu, Yaya Peng, Yong Zhao, Frank K. Soong:

Perceptual annotation of expressive speech. 46-51
Poster Session 1
- Karl Schnell, Arild Lacroix:

Joint analysis of speech frames for synthesis based on lossy tube models. 52-57 - Connie R. Adsett, Yannick Marchand:

Are rule-based syllabification methods adequate for languages with low syllabic complexity? the case of Italian. 58-63 - Mark A. Huckvale, Kayoko Yanagisawa:

Spoken language conversion with accent morphing. 64-70 - Grazyna Demenko, Agnieszka Wagner, Matthias Jilka, Bernd Möbius:

Comparative investigation of peak alignment in Polish and German unit selection corpora. 71-76 - Katarzyna Klessa, Marcin Szymanski, Stefan Breuer, Grazyna Demenko:

Optimization of Polish segmental duration prediction with CART. 77-80 - Toshio Hirai, Junichi Yamagishi, Seiichi Tenpaku:

Utilization of an HMM-based feature generation module in 5 ms segment concatenative speech synthesis. 81-84 - Damien Lolive, Nelly Barbot, Olivier Boëffard:

Clustering algorithm for F0 curves based on hidden Markov models. 85-89 - Rohit Kumar, Rashmi Gangadharaiah, Sharath Rao, Kishore Prahallad, Carolyn P. Rosé, Alan W. Black:

Building a better Indian English voice using "more data". 90-94 - Marc Schröder, Anna Hunecke:

Creating German unit selection voices for the MARY TTS platform from the BITS corpora. 95-100
Voice Conversion
- Kumi Ohta, Yamato Ohtani, Tomoki Toda, Hiroshi Saruwatari, Kiyohiro Shikano:

Regression approaches to voice quality controll based on one-to-many eigenvoice conversion. 101-106 - Daisuke Tani, Yamato Ohtani, Tomoki Toda, Hiroshi Saruwatari, Kiyohiro Shikano:

An evaluation of many-to-one voice conversion algorithms with pre-stored speaker data sets. 107-112 - João P. Cabral, Steve Renals, Korin Richmond, Junichi Yamagishi:

Towards an improved modeling of the glottal source in statistical parametric speech synthesis. 113-118 - Larbi Mesbahi, Vincent Barreaud, Olivier Boëffard:

GMM-based speech transformation systems under data reduction. 119-124
Speech Synthesis by HMM
- Junichi Yamagishi, Takao Kobayashi, Steve Renals, Simon King, Heiga Zen, Tomoki Toda, Keiichi Tokuda:

Improved average-voice-based speech synthesis using gender-mixed modeling and a parameter generation algorithm considering GV. 125-130 - Ranniery Maia, Tomoki Toda, Heiga Zen, Yoshihiko Nankaku, Keiichi Tokuda:

An excitation model for HMM-based speech synthesis based on residual modeling. 131-136 - Hui Liang, Yao Qian, Frank K. Soong:

An HMM-based bilingual (Mandarin-English) TTS. 137-142 - Justus C. Roux, Albert S. Visagie:

Data-driven approach to rapid prototyping Xhosa speech synthesis. 143-147
Tone and Tone Accent Languages
- Nobuaki Minematsu, Ryo Kuroiwa, Keikichi Hirose, Michiko Watanabe:

CRF-based statistical learning of Japanese accent sandhi for developing Japanese text-to-speech synthesis systems. 148-153 - Qinghua Sun, Keikichi Hirose, Nobuaki Minematsu:

Two-step generation of Mandarin F0 contours based on tone nucleus and superpositional models. 154-159 - Suphattharachai Chomphan, Takao Kobayashi:

Design of tree-based context clustering for an HMM-based Thai speech synthesis system. 160-165 - Arne Bachmann, Stefan Breuer:

Development of a BOSS unit selection module for tone languages. 166-171
Poster Session 2
- Alexander Kain, Jan P. H. van Santen:

Unit-selection text-to-speech synthesis using an asynchronous interpolation model. 172-177 - Ingo Hertrich, Hermann Ackermann:

Modelling voiceless speech segments by means of an additive procedure based on the computation of formant sinusoids. 178-181 - Arthur R. Toth, Alan W. Black:

Using articulatory position data in voice transformation. 182-187 - Anand Arokia Raj, Tanuja Sarkar, Sathish Chandra Pammi, Santhosh Yuvaraj, Mohit Bansal, Kishore Prahallad, Alan W. Black:

Text processing for text-to-speech systems in Indian languages. 188-193 - Daniel Erro, Asunción Moreno, Antonio Bonafonte:

Flexible harmonic/stochastic speech synthesis. 194-199 - Jan Romportl, Jirí Kala:

Prosody modelling in Czech text-to-speech synthesis. 200-205 - Yong Zhao, Chengsuo Zhang, Frank K. Soong, Min Chu, Xi Xiao:

Measuring attribute dissimilarity with HMM KL-divergence for speech synthesis. 206-210 - Jonathan Chevelu, Nelly Barbot, Olivier Boëffard, Arnaud Delhay:

Lagrangian relaxation for optimal corpus design. 211-216 - Aleksandra Krul, Géraldine Damnati, François Yvon, Cédric Boidin, Thierry Moudenc:

Adaptive database reduction for domain specific speech synthesis. 217-222 - Jordi Adell, Antonio Bonafonte, David Escudero:

Statistical analysis of filled pauses2 rhythm for disfluent speech synthesis. 223-227 - Wentao Gu, Tan Lee:

Quantitative analysis of F0 contours of emotional speech of Mandarin. 228-233
Prosody Modelling
- Slava Shechtman:

Maximum-likelihood dynamic intonation model for concatenative text-to-speech system. 234-239 - Uwe D. Reichel:

Data-driven extraction of intonation contour classes. 240-245 - Taniya Mishra, Emily Tucker Prud'hommeaux, Jan P. H. van Santen:

Word accentuation prediction using a neural net classifier. 246-251 - Leonardo Badino, Robert A. J. Clark:

Issues of optionality in pitch accent placement. 252-257
Inventory Construction
- Matthew P. Aylett, Simon King:

Single speaker segmentation and inventory selection using dynamic time warping self organization and joint multigram mapping. 258-263 - Tanya Lambert, Norbert Braunschweiler, Sabine Buchholz:

How (not) to select your voice corpus: random selection vs. phonologically balanced. 264-269 - Lukas Latacz, Yuk On Kong, Werner Verhelst:

Unit selection synthesis using long non-uniform units and phonemic identity matching. 270-275 - Martin Gruber, Daniel Tihelka, Jindrich Matousek:

Evaluation of various unit types in the unit selection approach for the Czech language using the Festival system. 276-281
Keynote 2
- Alan W. Black:

The Blizzard Challenge: evaluating corpus-based speech synthesis techniques.
Applications
- Donata Moers, Petra Wagner, Stefan Breuer:

Assessing the adequate treatment of fast speech in unit selection speech synthesis systems for the visually impaired. 282-287 - Maria K. Wolters, Pauline Campbell, Christine DePlacido, Amy Liddell, David Owens:

Making speech synthesis more accessible to older people. 288-293
Systems
- Heiga Zen, Takashi Nose, Junichi Yamagishi, Shinji Sako, Takashi Masuko, Alan W. Black, Keiichi Tokuda:

The HMM-based speech synthesis system (HTS) version 2.0. 294-299 - Christian Weiss, Luís C. Oliveira, Sérgio Paulo, Carlos Mendes, Luís Figueira, Marco Vala, Pedro Sequeira, Ana Paiva, Thurid Vogt, Elisabeth André:

eCIRCUS: building voices for autonomous speaking agents. 300-303 - Martin Barbisch, Grzegorz Dogil, Bernd Möbius, Bettina Säuberlich, Antje Schweitzer:

Unit selection synthesis in the Smartweb project. 304-309 - Hanna Silén, Elina Helander, Konsta Koppinen, Moncef Gabbouj:

Building a Finnish unit selection TTS system. 310-315
Poster Session 3
- Yannick Marchand, Connie R. Adsett, Robert I. Damper:

Evaluating automatic syllabification algorithms for English. 316-321 - John Kominek, Tanja Schultz, Alan W. Black:

Voice building from insufficient data - classroom experiences with web-based language development tools. 322-327 - Peter Cahill, Jan Macek, Julie Carson-Berndsen:

SVM based feature extraction in speech synthesis. 328-332 - Yoshihiko Nankaku, Kenichi Nakamura, Tomoki Toda, Keiichi Tokuda:

Spectral conversion based on statistical models including time-sequence matching. 333-338 - Esther Klabbers, Taniya Mishra, Jan P. H. van Santen:

Analysis of affective speech recordings using the superpositional intonation model. 339-344 - Sylvain Le Beux, Albert Rilliard, Christophe d'Alessandro:

Calliphony: a real-time intonation controller for expressive speech synthesis. 345-350 - Shyamal Kr. Das Mandal, Asoke Kumar Datta:

Epoch synchronous non-overlap-add (ESNOLA) method-based concatenative speech synthesis system for Bangla. 351-355 - Chatchawarn Hansakunbuntheung, Hiroaki Kato, Yoshinori Sagisaka:

Syllable-based Thai duration model using multi-level linear regression and syllable accommodation. 356-361 - Xavier Gonzalvo, Joan Claudi Socoró, Ignasi Iriondo Sanz, Carlos Monzo, Elisa Martínez:

Linguistic and mixed excitation improvements on a HMM-based speech synthesis for Castilian Spanish. 362-367 - Tetyana Lyudovyk, Valentyna Robeiko:

Inventory of intonation contours for text-to-speech synthesis. 368-373
Evaluation
- H. Timothy Bunnell, Jason Lilley:

Analysis methods for assessing TTS intelligibility. 374-379 - Brian Langner, Alan W. Black:

Understandable production of massive synthesis. 380-384 - Charlotte van Hooijdonk, Edwin Commandeur, Reinier Cozijn, Emiel Krahmer, Erwin Marsi:

The online evaluation of speech synthesis using eye movements. 385-390

manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.


Google
Google Scholar
Semantic Scholar
Internet Archive Scholar
CiteSeerX
ORCID














