BACK TO INDEX

Publications of Yue Wang at Hamarneh Lab
Articles in journal, book chapters
  1. Saurabh Garg, Haoyao Ruan, Ghassan Hamarneh, Dawn M. Behne, Allard Jongman, Joan Sereno, and Yue Wang. Mouth2Audio: Intelligible Audio Synthesis From Videos With Distinctive Vowel Articulation. International Journal of Speech Technology, 000(000):000-000, 2023. Keyword(s): Synthesis/Simulation/Augmentation, Tracking, Machine Learning, Deep Learning, Color/Multichannel/Vector-valued, Speech and Language, Facial Analysis, Spatio-Temporal. [bibtex-key = ijst2023b]


  2. Saurabh Garg, Lisa Y. W. Tang, Ghassan Hamarneh, Allard Jongman, Joan Sereno, and Yue Wang. Different facial cues for different speech styles in Mandarin tone articulation. Frontiers in Communication - Language Sciences, 8(1148240):1-19, 2023. Keyword(s): Tracking, Machine Learning, Color/Multichannel/Vector-valued, Speech and Language, Facial Analysis, Spatio-Temporal. [bibtex-key = frontcomm2023]


  3. Shubam Sachdeva, Haoyao Ruan, Ghassan Hamarneh, Dawn Behne, Allard Jongman, Joan Sereno, and Yue Wang. Plain-to-clear speech video conversion for enhanced intelligibility. International Journal of Speech Technology (IJST), pp 1-22, 2023. Keyword(s): Segmentation, Tracking, Color/Multichannel/Vector-valued, Speech and Language, Facial Analysis, Spatio-Temporal. [bibtex-key = ijst2023a]


  4. Saurabh Garg, Ghassan Hamarneh, Allard Jongman, Joan Sereno, and Yue Wang. ADFAC: Automatic Detection of Facial Articulatory Features. MethodX, 7:101006, 2020. Keyword(s): Segmentation, Tracking, Color/Multichannel/Vector-valued, Speech and Language, Facial Analysis, Spatio-Temporal, Software and Tools. [bibtex-key = methodx2020]


  5. Saurabh Garg, Lisa Y. W. Tang, Ghassan Hamarneh, Allard Jongman, Joan Sereno, and Yue Wang. Different facial cues for different speech styles in Mandarin tone articulation. The Journal of the Acoustical Society of America (JASA). Also appears in ASA Meeting, 148(4):2764-2764, 2020. Keyword(s): Tracking, Machine Learning, Color/Multichannel/Vector-valued, Speech and Language, Facial Analysis, Spatio-Temporal. [bibtex-key = jasa2020]


  6. Saurabh Garg, Ghassan Hamarneh, Allard Jongman, Joan Sereno, and Yue Wang. Computer-vision analysis reveals facial movements made during Mandarin tone production align with pitch trajectories. Speech Communications, 113:47-62, 2019. Keyword(s): Segmentation, Tracking, Color/Multichannel/Vector-valued, Speech and Language, Facial Analysis, Spatio-Temporal. [bibtex-key = speechcomm2019]


  7. Saurabh Garg, Lisa Y. W. Tang, Ghassan Hamarneh, Allard Jongman, Joan Sereno, and Yue Wang. Computer-vision analysis shows different facial movements for the production of different Mandarin tones. The Journal of the Acoustical Society of America (JASA). Also appears in ASA Meeting, 144(3):1720-1720, 2018. Keyword(s): Tracking, Machine Learning, Color/Multichannel/Vector-valued, Speech and Language, Facial Analysis, Spatio-Temporal. [bibtex-key = jasa2018b]


  8. Joan Sereno, Allard Jongman, Yue Wang, Ghassan Hamarneh, Lisa Y. W. Tang, Saurabh Garg, Paul Tupper, Bob McMurray, Charles Redmon, Yuyu Zeng, Beverly Hannah, Keith Leung, and Sylvia Cho. Linking production and perception of clear speech. The Journal of the Acoustical Society of America (JASA). Also appears in ASA Meeting, 144(3):1725-1725, 2018. Keyword(s): Tracking, Machine Learning, Color/Multichannel/Vector-valued, Speech and Language, Facial Analysis, Spatio-Temporal. [bibtex-key = jasa2018a]


  9. Lisa Y. W. Tang, Beverly Hannah, Allard Jongman, Joan Sereno, Yue Wang, and Ghassan Hamarneh. Examining visible articulatory features in clear and plain speech. Speech Communications, 75:1-13, 2015. Keyword(s): Segmentation, Tracking, Color/Multichannel/Vector-valued, Speech and Language, Facial Analysis, Spatio-Temporal. [bibtex-key = speechcomm2015]


Conference articles
  1. Saurabh Garg, Lisa Y. W. Tang, Ghassan Hamarneh, Allard Jongman, Joan Sereno, and Yue Wang. Different facial cues for different speech styles in Mandarin tone articulation. In Acoustical Society of America Meeting (ASA), pages 1, 2020. Keyword(s): Tracking, Machine Learning, Color/Multichannel/Vector-valued, Speech and Language, Facial Analysis, Spatio-Temporal. [bibtex-key = asa2020]


  2. Saurabh Garg, Ghassan Hamarneh, Allard Jongman, Joan Sereno, and Yue Wang. Joint Gender-, Tone-, Vowel- Classification via Novel Hierarchical Classification for Annotation of Monosyllabic Mandarin Word Tokens. In IEEE International Conference on Acoustics, Speech, and Signal Processing (IEEE ICASSP), pages 5744-5748, 2018. Keyword(s): Tracking, Segmentation, Color/Multichannel/Vector-valued, Speech and Language, Facial Analysis, Spatio-Temporal. [bibtex-key = icassp2018]


  3. Saurabh Garg, Lisa Y. W. Tang, Ghassan Hamarneh, Allard Jongman, Joan Sereno, and Yue Wang. Computer-vision analysis shows different facial movements for the production of different Mandarin tones. In Acoustical Society of America Meeting (ASA), pages 1, 2018. Keyword(s): Tracking, Machine Learning, Color/Multichannel/Vector-valued, Speech and Language, Facial Analysis, Spatio-Temporal. [bibtex-key = asa2018b]


  4. Yue Wang, Beverly Hannah, Keith Leung, Sylvia Cho, Joan Sereno, Allard Jongman, Charles Redmon, Yuyu Zeng, Ghassan Hamarneh, Lisa Y. W. Tang, Saurabh Garg, Paul Tupper, and Bob McMurray. Linking production and perception of clear speech. In Acoustical Society of America Meeting (ASA), pages 1, 2018. Keyword(s): Tracking, Machine Learning, Color/Multichannel/Vector-valued, Speech and Language, Facial Analysis, Spatio-Temporal. [bibtex-key = asa2018a]


  5. Lisa Y. W. Tang, Beverly Hannah, Allard Jongman, Joan Sereno, Yue Wang, and Ghassan Hamarneh. Examining Visible Articulatory Features in Clear and Conversational Speech. In The International Congress of Phonetic Sciences (ICPhS), pages 1-4, 2015. Keyword(s): Tracking, Segmentation, Color/Multichannel/Vector-valued, Speech and Language, Facial Analysis, Spatio-Temporal. [bibtex-key = icphs2015]


Internal reports
  1. Saurabh Garg, Haoyao Ruan, Ghassan Hamarneh, Dawn M. Behne, Allard Jongman, Joan Sereno, and Yue Wang. Mouth2audio: Intelligible Audio Synthesis from Videos with Distinctive Vowel Articulation. Technical report SSRN Electronic Journal 4146179, 6 2022. Keyword(s): Segmentation, Tracking, Color/Multichannel/Vector-valued, Speech and Language, Facial Analysis, Spatio-Temporal, Deep Learning. [bibtex-key = ssrn2022_4146179]



BACK TO INDEX




Disclaimer:

This material is presented to ensure timely dissemination of scholarly and technical work. Copyright and all rights therein are retained by authors or by other copyright holders. All person copying this information are expected to adhere to the terms and constraints invoked by each author's copyright. In most cases, these works may not be reposted without the explicit permission of the copyright holder.




Last modified: Thu Apr 11 13:23:22 2024
Author: hamarneh.


This document was translated from BibTEX by bibtex2html