Almeida, N., Silva, S., Teixeira, A. J. S., and Vieira, D.
(2016). Multi-device applications using the multi-
modal architecture. In Dahl, D., editor, Multimodal
Interaction with W3C Standards: Towards Natural
User Interfaces to Everything, (to appear). Springer,
New York, NY, USA.
Birkholz, P. (2013). Modeling consonant-vowel coarticu-
lation for articulatory speech synthesis. PLoS ONE,
8(4):1–17.
Browman, C. P. and Goldstein, L. (1990). Gestural speci-
fication using dynamically-defined articulatory struc-
tures. Journal of Phonetics, 18:299–320.
Cohen, M. M. and Massaro, D. W. (1993). Modeling coar-
ticulation in synthetic visual speech. In Models and
techniques in computer animation, pages 139–156.
Springer.
Files, B. T., Tjan, B. S., Jiang, J., and Bernstein, L. E.
(2015). Visual speech discrimination and identifica-
tion of natural and synthetic consonant stimuli. Fron-
tiers in psychology, 6.
Freitas, J., Candeias, S., Dias, M. S., Lleida, E., Ortega, A.,
Teixeira, A., Silva, S., Acarturk, C., and Orvalho, V.
(2014). The IRIS project: A liaison between industry
and academia towards natural multimodal communi-
cation. In Proc. Iberspeech, pages 338–347, Las Pal-
mas de Gran Can
´
aria, Spain.
Hall, N. (2010). Articulatory phonology. Language and
Linguistics Compass, 4(9):818–830.
Massaro, D. W. (2005). The Psychology and Technology of
Talking Heads: Applications in Language Learning,
pages 183–214. Springer Netherlands, Dordrecht.
Mattheyses, W. and Verhelst, W. (2015). Audiovisual
speech synthesis: An overview of the state-of-the-art.
Speech Communication, 66:182 – 217.
Nam, H., Goldstein, L., Browman, C., Rubin, P., Proctor,
M., and Saltzman, E. (2006). TADA manual. New
Haven, CT: Haskins Labs.
Oliveira, C. (2009). From Grapheme to Gesture. Linguis-
tic Contributions for an Articulatory Based Text-To-
Speech System. PhD thesis, University of Aveiro (in
Portuguese).
Rubin, P., Baer, T., and Mermelstein, P. (1981). An articula-
tory synthesizer for perceptual research. The Journal
of the Acoustical Society of America, 70(2):321–328.
Rubin, P., Saltzman, E., Goldstein, L., McGowan, R., Tiede,
M., and Browman, C. (1996). CASY and extensions
to the task-dynamic model. In Proc. Speech Prod.
Seminar, pages 125–128.
Saltzman, E. L. and Munhall, K. G. (1989). A dynamical
approach to gestural patterning in speech production.
Ecological psychology, 1(4):333–382.
Schabus, D., Pucher, M., and Hofer, G. (2014). Joint audio-
visual hidden semi-markov model-based speech syn-
thesis. J. of Selected Topics in Signal Proc., 8(2):336–
347.
Scott, A. D., Wylezinska, M., Birch, M. J., and Miquel,
M. E. (2014). Speech mri: Morphology and function.
Physica Medica, 30(6):604 – 618.
Serra, J., Ribeiro, M., Freitas, J., Orvalho, V., and Dias,
M. S. (2012). A proposal for a visual speech anima-
tion system for european portuguese. In Proc. Iber-
SPEECH, pages 267–276, Madrid, Spain. Springer.
Silva, S., Almeida, N., Pereira, C., Martins, A. I., Rosa,
A. F., e Silva, M. O., and Teixeira, A. (2015). Design
and development of multimodal applications: A vi-
sion on key issues and methods. In Proc. HCII, LNCS.
Teixeira, A., Oliveira, C., and Barbosa, P. (2008). European
Portuguese articulatory based text-to-speech: First re-
sults. In Proc. PROPOR, LNAI 5190 , pages 101–111.
Teixeira, A., Silva, L., Martinez, R., and Vaz, F. (2002).
SAPWindows - towards a versatile modular articu-
latory synthesizer. In Proc. of IEEE Workshop on
Speech Synthesis, pages 31–34.
Teixeira, A. J. S., Almeida, N., Pereira, C., e Silva, M. O.,
Vieira, D., and Silva, S. (2016). Applications of
the multimodal interaction architecture in ambient as-
sisted living. In Dahl, D., editor, Multimodal Inter-
action with W3C Standards: Towards Natural User
Interfaces to Everything, (to appear). Springer, New
York, NY, USA.
W3C Consortium (2003). W3C multimodal interaction
framework - technical note (accessed oct 2016).
ˇ
Zelezn
´
y, M., Kr
ˇ
noul, Z., and Jedli
ˇ
cka, P. (2015). Analy-
sis of Facial Motion Capture Data for Visual Speech
Synthesis, pages 81–88. Springer International Pub-
lishing, Cham.
BIOSIGNALS 2017 - 10th International Conference on Bio-inspired Systems and Signal Processing
172