default search action
Hung-yi Lee
Hung-Yi Lee
Person information
Other persons with a similar name
SPARQL queries
🛈 Please note that only 72% of the records listed on this page have a DOI. Therefore, DOI-based queries can only provide partial results.
Refine list
refinements active!
zoomed in on ?? of ?? records
view refined list in
export refined list as
showing all ?? records
2020 – today
- 2024
- [j27]Shu-Wen Yang, Heng-Jui Chang, Zili Huang, Andy T. Liu, Cheng-I Lai, Haibin Wu, Jiatong Shi, Xuankai Chang, Hsiang-Sheng Tsai, Wen-Chin Huang, Tzu-hsun Feng, Po-Han Chi, Yist Y. Lin, Yung-Sung Chuang, Tzu-Hsien Huang, Wei-Cheng Tseng, Kushal Lakhotia, Shang-Wen Li, Abdelrahman Mohamed, Shinji Watanabe, Hung-yi Lee:
A Large-Scale Evaluation of Speech Foundation Models. IEEE ACM Trans. Audio Speech Lang. Process. 32: 2884-2899 (2024) - [j26]Kai-Wei Chang, Haibin Wu, Yu-Kai Wang, Yuan-Kuei Wu, Hua Shen, Wei-Cheng Tseng, Iu-thing Kang, Shang-wen Li, Hung-Yi Lee:
SpeechPrompt: Prompting Speech Language Models for Speech Processing Tasks. IEEE ACM Trans. Audio Speech Lang. Process. 32: 3730-3744 (2024) - [j25]Shensian Syu, Juncheng Xie, Hung-yi Lee:
Improving Non-Autoregressive Translation Quality With Pretrained Language Model, Embedding Distillation and Upsampling Strategy for CTC. IEEE ACM Trans. Audio Speech Lang. Process. 32: 4121-4133 (2024) - [c215]Cheng-Han Chiang, Hung-yi Lee:
Merging Facts, Crafting Fallacies: Evaluating the Contradictory Nature of Aggregated Factual Claims in Long-Form Generations. ACL (Findings) 2024: 2734-2751 - [c214]Guan-Ting Lin, Cheng-Han Chiang, Hung-yi Lee:
Advancing Large Language Models to Capture Varied Speaking Styles and Respond Properly in Spoken Conversations. ACL (1) 2024: 6626-6642 - [c213]Haibin Wu, Ho-Lam Chung, Yi-Cheng Lin, Yuan-Kuei Wu, Xuanjun Chen, Yu-Chi Pai, Hsiu-Hsuan Wang, Kai-Wei Chang, Alexander H. Liu, Hung-yi Lee:
Codec-SUPERB: An In-Depth Analysis of Sound Codec Models. ACL (Findings) 2024: 10330-10348 - [c212]Shih-Cheng Huang, Pin-Zu Li, Yu-Chi Hsu, Kuang-Ming Chen, Yu-Tung Lin, Shih-Kai Hsiao, Richard Tzong-Han Tsai, Hung-yi Lee:
Chat Vector: A Simple Approach to Equip LLMs with Instruction Following and Model Alignment in New Languages. ACL (1) 2024: 10943-10959 - [c211]Siddhant Arora, Ankita Pasad, Chung-Ming Chien, Jionghao Han, Roshan S. Sharma, Jee-weon Jung, Hira Dhamyal, William Chen, Suwon Shon, Hung-yi Lee, Karen Livescu, Shinji Watanabe:
On the Evaluation of Speech Foundation Models for Spoken Language Understanding. ACL (Findings) 2024: 11923-11938 - [c210]Cheng-Han Chiang, Hung-yi Lee:
Over-Reasoning and Redundant Calculation of Large Language Models. EACL (2) 2024: 161-169 - [c209]Hsuan-Fu Wang, Yi-Jen Shih, Heng-Jui Chang, Layne Berry, Puyuan Peng, Hung-Yi Lee, Hsin-Min Wang, David Harwath:
SpeechCLIP+: Self-Supervised Multi-Task Representation Learning for Speech Via Clip and Speech-Image Data. ICASSP Workshops 2024: 465-469 - [c208]Fabian Ritter Gutierrez, Kuan-Po Huang, Dianwen Ng, Jeremy H. M. Wong, Hung-Yi Lee, Eng Siong Chng, Nancy F. Chen:
Noise Robust Distillation of Self-Supervised Speech Models via Correlation Metrics. ICASSP Workshops 2024: 495-499 - [c207]Chih-Kai Yang, Kuan-Po Huang, Ke-Han Lu, Chun-Yi Kuan, Chi-Yuan Hsiao, Hung-Yi Lee:
Investigating Zero-Shot Generalizability on Mandarin-English Code-Switched ASR And Speech-to-Text Translation of Recent Foundation Models with Self-Supervision and Weak Supervision. ICASSP Workshops 2024: 540-544 - [c206]Hung-Chieh Fang, Nai-Xuan Ye, Yi-Jen Shih, Puyuan Peng, Hsuan-Fu Wang, Layne Berry, Hung-Yi Lee, David Harwath:
Integrating Self-Supervised Speech Model with Pseudo Word-Level Targets from Visually-Grounded Speech Model. ICASSP Workshops 2024: 645-649 - [c205]Tzu-Han Lin, How-Shing Wang, Hao-Yung Weng, Kuang-Chen Peng, Zih-Ching Chen, Hung-Yi Lee:
PEFT for Speech: Unveiling Optimal Placement, Merging Strategies, and Ensemble Techniques. ICASSP Workshops 2024: 705-709 - [c204]Haibin Wu, Heng-Cheng Kuo, Yu Tsao, Hung-Yi Lee:
Scalable Ensemble-Based Detection Method Against Adversarial Attacks For Speaker Verification. ICASSP 2024: 4670-4674 - [c203]Yuan Tseng, Layne Berry, Yiting Chen, I-Hsiang Chiu, Hsuan-Hao Lin, Max Liu, Puyuan Peng, Yi-Jen Shih, Hung-Yu Wang, Haibin Wu, Poyao Huang, Chun-Mao Lai, Shang-Wen Li, David Harwath, Yu Tsao, Abdelrahman Mohamed, Chi-Luen Feng, Hung-Yi Lee:
AV-SUPERB: A Multi-Task Evaluation Benchmark for Audio-Visual Representation Models. ICASSP 2024: 6890-6894 - [c202]Xuanjun Chen, Haibin Wu, Chung-Che Wang, Hung-Yi Lee, Jyh-Shing Roger Jang:
Multimodal Transformer Distillation for Audio-Visual Synchronization. ICASSP 2024: 7755-7759 - [c201]Kuan-Po Huang, Chih-Kai Yang, Yu-Kuan Fu, Ewan Dunbar, Hung-Yi Lee:
Zero Resource Code-Switched Speech Benchmark Using Speech Utterance Pairs for Multiple Spoken Languages. ICASSP 2024: 10006-10010 - [c200]Guan-Ting Lin, Prashanth Gurunath Shivakumar, Ankur Gandhe, Chao-Han Huck Yang, Yile Gu, Shalini Ghosh, Andreas Stolcke, Hung-Yi Lee, Ivan Bulyko:
Paralinguistics-Enhanced Large Language Modeling of Spoken Dialogue. ICASSP 2024: 10316-10320 - [c199]Chien-Yu Huang, Ke-Han Lu, Shih-Heng Wang, Chi-Yuan Hsiao, Chun-Yi Kuan, Haibin Wu, Siddhant Arora, Kai-Wei Chang, Jiatong Shi, Yifan Peng, Roshan S. Sharma, Shinji Watanabe, Bhiksha Ramakrishnan, Shady Shehata, Hung-Yi Lee:
Dynamic-Superb: Towards a Dynamic, Collaborative, and Comprehensive Instruction-Tuning Benchmark For Speech. ICASSP 2024: 12136-12140 - [c198]Chyi-Jiunn Lin, Guan-Ting Lin, Yung-Sung Chuang, Wei-Lun Wu, Shang-Wen Li, Abdelrahman Mohamed, Hung-Yi Lee, Lin-Shan Lee:
SpeechDPR: End-To-End Spoken Passage Retrieval For Open-Domain Spoken Question Answering. ICASSP 2024: 12476-12480 - [c197]Kevin Everson, Yile Gu, Chao-Han Huck Yang, Prashanth Gurunath Shivakumar, Guan-Ting Lin, Jari Kolehmainen, Ivan Bulyko, Ankur Gandhe, Shalini Ghosh, Wael Hamza, Hung-Yi Lee, Ariya Rastrow, Andreas Stolcke:
Towards ASR Robust Spoken Language Understanding Through in-Context Learning with Word Confusion Networks. ICASSP 2024: 12856-12860 - [c196]Shih-Cheng Huang, Shih-Heng Wang, Min-Han Shih, Saurav Sahay, Hung-yi Lee:
Systematic Analysis for Pretrained Language Model Priming for Parameter-Efficient Fine-tuning. NAACL (Student Research Workshop) 2024: 1-7 - [i271]Chih-Kai Yang, Kuan-Po Huang, Ke-Han Lu, Chun-Yi Kuan, Chi-Yuan Hsiao, Hung-yi Lee:
Investigating Zero-Shot Generalizability on Mandarin-English Code-Switched ASR and Speech-to-text Translation of Recent Foundation Models with Self-Supervision and Weak Supervision. CoRR abs/2401.00273 (2024) - [i270]Tzu-Han Lin, How-Shing Wang, Hao-Yung Weng, Kuang-Chen Peng, Zih-Ching Chen, Hung-yi Lee:
PEFT for Speech: Unveiling Optimal Placement, Merging Strategies, and Ensemble Techniques. CoRR abs/2401.02122 (2024) - [i269]Kevin Everson, Yile Gu, Chao-Han Huck Yang, Prashanth Gurunath Shivakumar, Guan-Ting Lin, Jari Kolehmainen, Ivan Bulyko, Ankur Gandhe, Shalini Ghosh, Wael Hamza, Hung-yi Lee, Ariya Rastrow, Andreas Stolcke:
Towards ASR Robust Spoken Language Understanding Through In-Context Learning With Word Confusion Networks. CoRR abs/2401.02921 (2024) - [i268]Chen-An Li, Hung-Yi Lee:
Examining Forgetting in Continual Pre-training of Aligned Large Language Models. CoRR abs/2401.03129 (2024) - [i267]David Cheng-Han Chiang, Hung-Yi Lee:
Over-Reasoning and Redundant Calculation of Large Language Models. CoRR abs/2401.11467 (2024) - [i266]Chyi-Jiunn Lin, Guan-Ting Lin, Yung-Sung Chuang, Wei-Lun Wu, Shang-Wen Li, Abdelrahman Mohamed, Hung-yi Lee, Lin-Shan Lee:
SpeechDPR: End-to-End Spoken Passage Retrieval for Open-Domain Spoken Question Answering. CoRR abs/2401.13463 (2024) - [i265]Wei-Ping Huang, Sung-Feng Huang, Hung-yi Lee:
Maximizing Data Efficiency for Cross-Lingual TTS Adaptation by Self-Supervised Representation Mixing and Embedding Initialization. CoRR abs/2402.01692 (2024) - [i264]Liang-Hsuan Tseng, En-Pei Hu, David Cheng-Han Chiang, Yuan Tseng, Hung-yi Lee, Lin-Shan Lee, Shao-Hua Sun:
REBORN: Reinforcement-Learned Boundary Segmentation with Iterative Training for Unsupervised ASR. CoRR abs/2402.03988 (2024) - [i263]David Cheng-Han Chiang, Hung-yi Lee:
Merging Facts, Crafting Fallacies: Evaluating the Contradictory Nature of Aggregated Factual Claims in Long-Form Generations. CoRR abs/2402.05629 (2024) - [i262]Hung-Chieh Fang, Nai-Xuan Ye, Yi-Jen Shih, Puyuan Peng, Hsuan-Fu Wang, Layne Berry, Hung-yi Lee, David Harwath:
Integrating Self-supervised Speech Model with Pseudo Word-level Targets from Visually-grounded Speech Model. CoRR abs/2402.05819 (2024) - [i261]Hsuan-Fu Wang, Yi-Jen Shih, Heng-Jui Chang, Layne Berry, Puyuan Peng, Hung-yi Lee, Hsin-Min Wang, David Harwath:
SpeechCLIP+: Self-supervised multi-task representation learning for speech via CLIP and speech-image data. CoRR abs/2402.06959 (2024) - [i260]Guan-Ting Lin, David Cheng-Han Chiang, Hung-yi Lee:
Advancing Large Language Models to Capture Varied Speaking Styles and Respond Properly in Spoken Conversations. CoRR abs/2402.12786 (2024) - [i259]Haibin Wu, Huang-Cheng Chou, Kai-Wei Chang, Lucas Goncalves, Jiawei Du, Jyh-Shing Roger Jang, Chi-Chun Lee, Hung-Yi Lee:
EMO-SUPERB: An In-depth Look at Speech Emotion Recognition. CoRR abs/2402.13018 (2024) - [i258]Haibin Wu, Ho-Lam Chung, Yi-Cheng Lin, Yuan-Kuei Wu, Xuanjun Chen, Yu-Chi Pai, Hsiu-Hsuan Wang, Kai-Wei Chang, Alexander H. Liu, Hung-yi Lee:
Codec-SUPERB: An In-Depth Analysis of Sound Codec Models. CoRR abs/2402.13071 (2024) - [i257]Haibin Wu, Xuanjun Chen, Yi-Cheng Lin, Kai-Wei Chang, Ho-Lam Chung, Alexander H. Liu, Hung-yi Lee:
Towards audio language modeling - an overview. CoRR abs/2402.13236 (2024) - [i256]Shu-Wen Yang, Heng-Jui Chang, Zili Huang, Andy T. Liu, Cheng-I Lai, Haibin Wu, Jiatong Shi, Xuankai Chang, Hsiang-Sheng Tsai, Wen-Chin Huang, Tzu-hsun Feng, Po-Han Chi, Yist Y. Lin, Yung-Sung Chuang, Tzu-Hsien Huang, Wei-Cheng Tseng, Kushal Lakhotia, Shang-Wen Li, Abdelrahman Mohamed, Shinji Watanabe, Hung-yi Lee:
A Large-Scale Evaluation of Speech Foundation Models. CoRR abs/2404.09385 (2024) - [i255]Li-Chun Lu, Shou-Jen Chen, Tsung-Min Pai, Chan-Hung Yu, Hung-yi Lee, Shao-Hua Sun:
LLM Discussion: Enhancing the Creativity of Large Language Models via Discussion Framework and Role-Play. CoRR abs/2405.06373 (2024) - [i254]Kuang-Ming Chen, Hung-yi Lee:
InstructionCP: A fast approach to transfer Large Language Models into target language. CoRR abs/2405.20175 (2024) - [i253]Hsuan Su, Hua Farn, Fan-Yun Sun, Shang-Tse Chen, Hung-yi Lee:
Task Arithmetic can Mitigate Synthetic-to-Real Gap in Automatic Speech Recognition. CoRR abs/2406.02925 (2024) - [i252]Fabian Ritter Gutierrez, Kuan-Po Huang, Jeremy H. M. Wong, Dianwen Ng, Hung-yi Lee, Nancy F. Chen, Eng Siong Chng:
Dataset-Distillation Generative Model for Speech Emotion Recognition. CoRR abs/2406.02963 (2024) - [i251]Xuanjun Chen, Jiawei Du, Haibin Wu, Jyh-Shing Roger Jang, Hung-yi Lee:
Neural Codec-based Adversarial Sample Detection for Speaker Verification. CoRR abs/2406.04582 (2024) - [i250]Yi-Cheng Lin, Tzu-Quan Lin, Hsi-Che Lin, Andy T. Liu, Hung-yi Lee:
On the social bias of speech self-supervised models. CoRR abs/2406.04997 (2024) - [i249]Tzu-Quan Lin, Hung-yi Lee, Hao Tang:
DAISY: Data Adaptive Self-Supervised Early Exit for Speech Representation Models. CoRR abs/2406.05464 (2024) - [i248]Chih-Kai Yang, Kuan-Po Huang, Hung-yi Lee:
Do Prompts Really Prompt? Exploring the Prompt Understanding Capability of Whisper. CoRR abs/2406.05806 (2024) - [i247]Haibin Wu, Yuan Tseng, Hung-yi Lee:
CodecFake: Enhancing Anti-Spoofing Models Against Deepfake Audios from Codec-Based Speech Synthesis Systems. CoRR abs/2406.07237 (2024) - [i246]Chun-Yi Kuan, Wei-Ping Huang, Hung-yi Lee:
Understanding Sounds, Missing the Questions: The Challenge of Object Hallucination in Large Audio-Language Models. CoRR abs/2406.08402 (2024) - [i245]Jiatong Shi, Shih-Heng Wang, William Chen, Martijn Bartelds, Vanya Bannihatti Kumar, Jinchuan Tian, Xuankai Chang, Dan Jurafsky, Karen Livescu, Hung-yi Lee, Shinji Watanabe:
ML-SUPERB 2.0: Benchmarking Multilingual Speech Models Across Modeling Constraints, Languages, and Datasets. CoRR abs/2406.08641 (2024) - [i244]Cheng-Kuang Wu, Zhi Rui Tam, Chieh-Yen Lin, Yun-Nung Chen, Hung-yi Lee:
StreamBench: Towards Benchmarking Continuous Improvement of Language Agents. CoRR abs/2406.08747 (2024) - [i243]Siddhant Arora, Ankita Pasad, Chung-Ming Chien, Jionghao Han, Roshan S. Sharma, Jee-weon Jung, Hira Dhamyal, William Chen, Suwon Shon, Hung-yi Lee, Karen Livescu, Shinji Watanabe:
On the Evaluation of Speech Foundation Models for Spoken Language Understanding. CoRR abs/2406.10083 (2024) - [i242]Hung-Ting Su, Chun-Tong Chao, Ya-Ching Hsu, Xudong Lin, Yulei Niu, Hung-Yi Lee, Winston H. Hsu:
Investigating Video Reasoning Capability of Large Language Models with Tropes in Movies. CoRR abs/2406.10923 (2024) - [i241]Guan-Ting Lin, Wei-Ping Huang, Hung-yi Lee:
Continual Test-time Adaptation for End-to-end Speech Recognition on Noisy Speech. CoRR abs/2406.11064 (2024) - [i240]Guan-Ting Lin, Hung-yi Lee:
Can LLMs Understand the Implication of Emphasized Sentences in Dialogue? CoRR abs/2406.11065 (2024) - [i239]Ke-Han Lu, Zhehuai Chen, Szu-Wei Fu, He Huang, Boris Ginsburg, Yu-Chiang Frank Wang, Hung-yi Lee:
DeSTA: Enhancing Speech Language Models through Descriptive Speech-Text Alignment. CoRR abs/2406.18871 (2024) - [i238]Tzu-Han Lin, Chen-An Li, Hung-yi Lee, Yun-Nung Chen:
DogeRM: Equipping Reward Models with Domain Knowledge through Model Merging. CoRR abs/2407.01470 (2024) - [i237]Yu-Kuan Fu, Cheng-Kuang Lee, Hsiu-Hsuan Wang, Hung-yi Lee:
Investigating the Effects of Large-Scale Pseudo-Stereo Data and Different Speech Foundation Model on Dialogue Generative Spoken Language Model. CoRR abs/2407.01911 (2024) - [i236]Cheng-Han Chiang, Wei-Chih Chen, Chun-Yi Kuan, Chienchou Yang, Hung-yi Lee:
Large Language Model as an Assignment Evaluator: Insights, Feedback, and Challenges in a 1000+ Student Course. CoRR abs/2407.05216 (2024) - [i235]Yi-Cheng Lin, Tzu-Quan Lin, Chih-Kai Yang, Ke-Han Lu, Wei-Chih Chen, Chun-Yi Kuan, Hung-yi Lee:
Listen and Speak Fairly: A Study on Semantic Gender Bias in Speech Integrated Large Language Models. CoRR abs/2407.06957 (2024) - [i234]Chun-Yi Kuan, Chih-Kai Yang, Wei-Ping Huang, Ke-Han Lu, Hung-yi Lee:
Speech-Copilot: Leveraging Large Language Models for Speech Processing via Task Decomposition, Modularization, and Program Generation. CoRR abs/2407.09886 (2024) - [i233]Liang-Hsuan Tseng, Zih-Ching Chen, Wei-Shun Chang, Cheng-Kuang Lee, Tsung-Ren Huang, Hung-yi Lee:
Leave No Knowledge Behind During Knowledge Distillation: Towards Practical and Effective Knowledge Distillation for Code-Switching ASR Using Realistic Data. CoRR abs/2407.10603 (2024) - [i232]Cheng-Kuang Wu, Zhi Rui Tam, Chao-Chung Wu, Chieh-Yen Lin, Hung-yi Lee, Yun-Nung Chen:
I Need Help! Evaluating LLM's Ability to Ask for Users' Support: A Case Study on Text-to-SQL Generation. CoRR abs/2407.14767 (2024) - [i231]Wenze Ren, Yi-Cheng Lin, Huang-Cheng Chou, Haibin Wu, Yi-Chiao Wu, Chi-Chun Lee, Hung-yi Lee, Yu Tsao:
EMO-Codec: An In-Depth Look at Emotion Preservation capacity of Legacy and Neural Codec Models With Subjective and Objective Evaluations. CoRR abs/2407.15458 (2024) - [i230]Zhi Rui Tam, Cheng-Kuang Wu, Yi-Lin Tsai, Chieh-Yen Lin, Hung-yi Lee, Yun-Nung Chen:
Let Me Speak Freely? A Study on the Impact of Format Restrictions on Performance of Large Language Models. CoRR abs/2408.02442 (2024) - [i229]Shachi H. Kumar, Saurav Sahay, Sahisnu Mazumder, Eda Okur, Ramesh Manuvinakurike, Nicole Beckage, Hsuan Su, Hung-yi Lee, Lama Nachman:
Decoding Biases: Automated Methods and LLM Judges for Gender Bias Detection in Language Models. CoRR abs/2408.03907 (2024) - [i228]Yi-Cheng Lin, Wei-Chih Chen, Hung-yi Lee:
Spoken Stereoset: On Evaluating Social Bias Toward Speaker in Speech Large Language Models. CoRR abs/2408.07665 (2024) - [i227]Kai-Wei Chang, Haibin Wu, Yu-Kai Wang, Yuan-Kuei Wu, Hua Shen, Wei-Cheng Tseng, Iu-thing Kang, Shang-Wen Li, Hung-yi Lee:
SpeechPrompt: Prompting Speech Language Models for Speech Processing Tasks. CoRR abs/2408.13040 (2024) - [i226]Chien-yu Huang, Min-Han Shih, Ke-Han Lu, Chi-Yuan Hsiao, Hung-yi Lee:
SpeechCaps: Advancing Instruction-Based Universal Speech Models with Multi-Talker Speaking Style Captioning. CoRR abs/2408.13891 (2024) - [i225]Juncheng Xie, Shensian Syu, Hung-yi Lee:
Non-instructional Fine-tuning: Enabling Instruction-Following Capabilities in Pre-trained Language Models without Instruction-Following Data. CoRR abs/2409.00096 (2024) - [i224]Tzu-Quan Lin, Guan-Ting Lin, Hung-yi Lee, Hao Tang:
Property Neurons in Self-Supervised Speech Transformers. CoRR abs/2409.05910 (2024) - [i223]Jiawei Du, I-Ming Lin, I-Hsiang Chiu, Xuanjun Chen, Haibin Wu, Wenze Ren, Yu Tsao, Hung-yi Lee, Jyh-Shing Roger Jang:
DFADD: The Diffusion and Flow-Matching Based Audio Deepfake Dataset. CoRR abs/2409.08731 (2024) - [i222]Ming-Hao Hsu, Kuan Po Huang, Hung-yi Lee:
Meta-Whisper: Speech-Based Meta-ICL for ASR on Low-Resource Languages. CoRR abs/2409.10429 (2024) - [i221]Hsi-Che Lin, Yi-Cheng Lin, Huang-Cheng Chou, Hung-yi Lee:
Improving Speech Emotion Recognition in Under-Resourced Languages via Speech-to-Speech Translation with Bootstrapping Data Selection. CoRR abs/2409.10985 (2024) - [i220]Haibin Wu, Xuanjun Chen, Yi-Cheng Lin, Kai-Wei Chang, Jiawei Du, Ke-Han Lu, Alexander H. Liu, Ho-Lam Chung, Yuan-Kuei Wu, Dongchao Yang, Songxiang Liu, Yi-Chiao Wu, Xu Tan, James R. Glass, Shinji Watanabe, Hung-yi Lee:
Codec-SUPERB @ SLT 2024: A lightweight benchmark for neural audio codec models. CoRR abs/2409.14085 (2024) - [i219]Hung-Ting Su, Ya-Ching Hsu, Xudong Lin, Xiang Qian Shi, Yulei Niu, Han-Yuan Hsu, Hung-yi Lee, Winston H. Hsu:
Unveiling Narrative Reasoning Limits of Large Language Models with Trope in Movie Synopses. CoRR abs/2409.14324 (2024) - [i218]Andy T. Liu, Yi-Cheng Lin, Haibin Wu, Stefan Winkler, Hung-yi Lee:
Efficient Training of Self-Supervised Speech Foundation Models on a Compute Budget. CoRR abs/2409.16295 (2024) - [i217]Ke-Han Lu, Zhehuai Chen, Szu-Wei Fu, Chao-Han Huck Yang, Jagadeesh Balam, Boris Ginsburg, Yu-Chiang Frank Wang, Hung-yi Lee:
Developing Instruction-Following Speech Language Model Without Speech Instruction-Tuning Data. CoRR abs/2409.20007 (2024) - 2023
- [j24]Po-Chun Hsu, Da-Rong Liu, Andy T. Liu, Hung-yi Lee:
Parallel Synthesis for Autoregressive Speech Generation. IEEE ACM Trans. Audio Speech Lang. Process. 31: 3095-3111 (2023) - [j23]Yun-Yen Chuang, Hung-Min Hsu, Kevin Lin, Ray-I Chang, Hung-Yi Lee:
MetaEx-GAN: Meta Exploration to Improve Natural Language Generation via Generative Adversarial Networks. IEEE ACM Trans. Audio Speech Lang. Process. 31: 3968-3980 (2023) - [c195]David Cheng-Han Chiang, Hung-yi Lee:
Are Synonym Substitution Attacks Really Synonym Substitution Attacks? ACL (Findings) 2023: 1853-1878 - [c194]Suwon Shon, Siddhant Arora, Chyi-Jiunn Lin, Ankita Pasad, Felix Wu, Roshan S. Sharma, Wei-Lun Wu, Hung-yi Lee, Karen Livescu, Shinji Watanabe:
SLUE Phase-2: A Benchmark Suite of Diverse Spoken Language Understanding Tasks. ACL (1) 2023: 8906-8937 - [c193]Derek Xu, Shuyan Dong, Changhan Wang, Suyoun Kim, Zhaojiang Lin, Bing Liu, Akshat Shrivastava, Shang-Wen Li, Liang-Hsuan Tseng, Guan-Ting Lin, Alexei Baevski, Hung-yi Lee, Yizhou Sun, Wei Wang:
Introducing Semantics into Speech Encoders. ACL (1) 2023: 11413-11429 - [c192]David Cheng-Han Chiang, Hung-yi Lee:
Can Large Language Models Be an Alternative to Human Evaluations? ACL (1) 2023: 15607-15631 - [c191]Kai-Wei Chang, Ming-Hsin Chen, Yun-Ping Lin, Jing Neng Hsu, Paul Kuo-Ming Huang, Chien-Yu Huang, Shang-Wen Li, Hung-Yi Lee:
Prompting and Adapter Tuning For Self-Supervised Encoder-Decoder Speech Model. ASRU 2023: 1-8 - [c190]Wei-Ping Huang, Sung-Feng Huang, Hung-Yi Lee:
Maximizing Data Efficiency for Cross-Lingual TTS Adaptation by Self-Supervised Representation Mixing and Embedding Initialization. ASRU 2023: 1-8 - [c189]Chun-Yi Kuan, Chen-An Li, Tsu-Yuan Hsu, Tse-Yang Lin, Ho-Lam Chung, Kai-Wei Chang, Shuo-Yiin Chang, Hung-Yi Lee:
Towards General-Purpose Text-Instruction-Guided Voice Conversion. ASRU 2023: 1-8 - [c188]Tzu-Quan Lin, Hung-Yi Lee, Hao Tang:
MelHuBERT: A Simplified Hubert on Mel Spectrograms. ASRU 2023: 1-8 - [c187]Jiatong Shi, William Chen, Dan Berrebbi, Hsiu-Hsuan Wang, Wei-Ping Huang, En-Pei Hu, Ho-Lam Chuang, Xuankai Chang, Yuxun Tang, Shang-Wen Li, Abdelrahman Mohamed, Hung-Yi Lee, Shinji Watanabe:
Findings of the 2023 ML-Superb Challenge: Pre-Training And Evaluation Over More Languages And Beyond. ASRU 2023: 1-8 - [c186]Yu-Hsiang Wang, Huang-Yu Chen, Kai-Wei Chang, Winston H. Hsu, Hung-Yi Lee:
Minisuperb: Lightweight Benchmark for Self-Supervised Speech Models. ASRU 2023: 1-8 - [c185]Jun-You Wang, Hung-Yi Lee, Jyh-Shing Roger Jang, Li Su:
Zero-Shot Singing Voice Synthesis from Musical Score. ASRU 2023: 1-8 - [c184]Haibin Wu, Jiawen Kang, Lingwei Meng, Helen Meng, Hung-yi Lee:
The Defender's Perspective on Automatic Speaker Verification: An Overview. DADA@IJCAI 2023: 6-11 - [c183]David Cheng-Han Chiang, Hung-yi Lee:
A Closer Look into Using Large Language Models for Automatic Evaluation. EMNLP (Findings) 2023: 8928-8942 - [c182]Layne Berry, Yi-Jen Shih, Hsuan-Fu Wang, Heng-Jui Chang, Hung-Yi Lee, David Harwath:
M-SpeechCLIP: Leveraging Large-Scale, Pre-Trained Models for Multilingual Speech to Image Retrieval. ICASSP 2023: 1-5 - [c181]Hsuan-Jui Chen, Yen Meng, Hung-yi Lee:
Once-for-All Sequence Compression for Self-Supervised Speech Models. ICASSP 2023: 1-5 - [c180]Zih-Ching Chen, Yu-Shun Sung, Hung-Yi Lee:
Chapter: Exploiting Convolutional Neural Network Adapters for Self-Supervised Speech Models. ICASSP Workshops 2023: 1-5 - [c179]Dongji Gao, Jiatong Shi, Shun-Po Chuang, Leibny Paola García, Hung-Yi Lee, Shinji Watanabe, Sanjeev Khudanpur:
Euro: Espnet Unsupervised ASR Open-Source Toolkit. ICASSP 2023: 1-5 - [c178]Chan-Jan Hsu, Ho-Lam Chung, Hung-Yi Lee, Yu Tsao:
T5lephone: Bridging Speech and Text Self-Supervised Models for Spoken Language Understanding Via Phoneme Level T5. ICASSP 2023: 1-5 - [c177]Sung-Feng Huang, Chia-Ping Chen, Zhi-Sheng Chen, Yu-Pao Tsai, Hung-Yi Lee:
Personalized Lightweight Text-to-Speech: Voice Cloning with Adaptive Structured Pruning. ICASSP 2023: 1-5 - [c176]Kuan-Po Huang, Tzu-hsun Feng, Yu-Kuan Fu, Tsu-Yuan Hsu, Po-Chieh Yen, Wei-Cheng Tseng, Kai-Wei Chang, Hung-Yi Lee:
Ensemble Knowledge Distillation of Self-Supervised Speech Models. ICASSP 2023: 1-5 - [c175]Jiatong Shi, Chan-Jan Hsu, Ho-Lam Chung, Dongji Gao, Paola García, Shinji Watanabe, Ann Lee, Hung-Yi Lee:
Bridging Speech and Textual Pre-Trained Models With Unsupervised ASR. ICASSP 2023: 1-5 - [c174]Yuan Tseng, Cheng-I Jeff Lai, Hung-Yi Lee:
Cascading and Direct Approaches to Unsupervised Constituency Parsing on Spoken Sentences. ICASSP 2023: 1-5 - [c173]Guan-Ting Liu, En-Pei Hu, Pu-Jen Cheng, Hung-Yi Lee, Shao-Hua Sun:
Hierarchical Programmatic Reinforcement Learning via Learning to Compose Programs. ICML 2023: 21672-21697 - [c172]Zih-Ching Chen, Chao-Han Huck Yang, Bo Li, Yu Zhang, Nanxin Chen, Shuo-Yiin Chang, Rohit Prabhavalkar, Hung-yi Lee, Tara N. Sainath:
How to Estimate Model Transferability of Pre-Trained Speech Models? INTERSPEECH 2023: 456-460 - [c171]Jiatong Shi, Dan Berrebbi, William Chen, En-Pei Hu, Wei-Ping Huang, Ho-Lam Chung, Xuankai Chang, Shang-Wen Li, Abdelrahman Mohamed, Hung-yi Lee, Shinji Watanabe:
ML-SUPERB: Multilingual Speech Universal PERformance Benchmark. INTERSPEECH 2023: 884-888 - [c170]Guan-Wei Wu, Guan-Ting Lin, Shang-Wen Li, Hung-yi Lee:
Improving Textless Spoken Language Understanding with Discrete Units as Intermediate Target. INTERSPEECH 2023: 1503-1507 - [c169]Cheng-Han Chiang, Wei-Ping Huang, Hung-yi Lee:
Why We Should Report the Details in Subjective Evaluation of TTS More Rigorously. INTERSPEECH 2023: 5551-5555 - [c168]David Cheng-Han Chiang, Hung-yi Lee, Yung-Sung Chuang, James R. Glass:
Revealing the Blind Spot of Sentence Encoder Evaluation by HEROS. RepL4NLP@ACL 2023: 289-302 - [e2]C. Maria Keet, Hung-Yi Lee, Sina Zarrieß:
Proceedings of the 16th International Natural Language Generation Conference, INLG 2023, Prague, Czechia, September 11 - 15, 2023. Association for Computational Linguistics 2023, ISBN 979-8-89176-001-1 [contents] - [i216]Guan-Ting Liu, En-Pei Hu, Pu-Jen Cheng, Hung-Yi Lee, Shao-Hua Sun:
Hierarchical Programmatic Reinforcement Learning via Learning to Compose Programs. CoRR abs/2301.12950 (2023) - [i215]Hsuan Su, Shachi H. Kumar, Sahisnu Mazumder, Wenda Chen, Ramesh Manuvinakurike, Eda Okur, Saurav Sahay, Lama Nachman, Shang-Tse Chen, Hung-yi Lee:
Position Matters! Empirical Study of Order Effect in Knowledge-grounded Dialogue. CoRR abs/2302.05888 (2023) - [i214]Kuan-Po Huang, Tzu-hsun Feng, Yu-Kuan Fu, Tsu-Yuan Hsu, Po-Chieh Yen, Wei-Cheng Tseng, Kai-Wei Chang, Hung-yi Lee:
Ensemble knowledge distillation of self-supervised speech models. CoRR abs/2302.12757 (2023) - [i213]Kai-Wei Chang, Yu-Kai Wang, Hua Shen, Iu-thing Kang, Wei-Cheng Tseng, Shang-Wen Li, Hung-yi Lee:
SpeechPrompt v2: Prompt Tuning for Speech Classification Tasks. CoRR abs/2303.00733 (2023) - [i212]Yuan Tseng, Cheng-I Lai, Hung-yi Lee:
Cascading and Direct Approaches to Unsupervised Constituency Parsing on Spoken Sentences. CoRR abs/2303.08809 (2023) - [i211]Sung-Feng Huang, Chia-Ping Chen, Zhi-Sheng Chen, Yu-Pao Tsai, Hung-yi Lee:
Personalized Lightweight Text-to-Speech: Voice Cloning with Adaptive Structured Pruning. CoRR abs/2303.11816 (2023) - [i210]David Cheng-Han Chiang, Hung-yi Lee:
Can Large Language Models Be an Alternative to Human Evaluations? CoRR abs/2305.01937 (2023) - [i209]Yu-Kuan Fu, Liang-Hsuan Tseng, Jiatong Shi, Chen-An Li, Tsu-Yuan Hsu, Shinji Watanabe, Hung-Yi Lee:
Improving Cascaded Unsupervised Speech Translation with Denoising Back-translation. CoRR abs/2305.07455 (2023) - [i208]Jiatong Shi, Dan Berrebbi, William Chen, Ho-Lam Chung, En-Pei Hu, Wei-Ping Huang, Xuankai Chang, Shang-Wen Li, Abdelrahman Mohamed, Hung-yi Lee, Shinji Watanabe:
ML-SUPERB: Multilingual Speech Universal PERformance Benchmark. CoRR abs/2305.10615 (2023) - [i207]Haibin Wu, Jiawen Kang, Lingwei Meng, Helen Meng, Hung-yi Lee:
The defender's perspective on automatic speaker verification: An overview. CoRR abs/2305.12804 (2023) - [i206]Guan-Wei Wu, Guan-Ting Lin, Shang-Wen Li, Hung-yi Lee:
Improving Textless Spoken Language Understanding with Discrete Units as Intermediate Target. CoRR abs/2305.18096 (2023) - [i205]Yu-Hsiang Wang, Huang-Yu Chen, Kai-Wei Chang, Winston H. Hsu, Hung-yi Lee:
MiniSUPERB: Lightweight Benchmark for Self-supervised Speech Models. CoRR abs/2305.19011 (2023) - [i204]Zih-Ching Chen, Chao-Han Huck Yang, Bo Li, Yu Zhang, Nanxin Chen, Shuo-Yiin Chang, Rohit Prabhavalkar, Hung-yi Lee, Tara N. Sainath:
How to Estimate Model Transferability of Pre-Trained Speech Models? CoRR abs/2306.01015 (2023) - [i203]Haibin Wu, Kai-Wei Chang, Yuan-Kuei Wu, Hung-yi Lee:
SpeechGen: Unlocking the Generative Power of Speech Language Models with Prompts. CoRR abs/2306.02207 (2023) - [i202]David Cheng-Han Chiang, Yung-Sung Chuang, James R. Glass, Hung-yi Lee:
Revealing the Blind Spot of Sentence Encoder Evaluation by HEROS. CoRR abs/2306.05083 (2023) - [i201]Shen-sian Syu, Juncheng Xie, Hung-yi Lee:
Improving Non-autoregressive Translation Quality with Pretrained Language Model, Embedding Distillation and Upsampling Strategy for CTC. CoRR abs/2306.06345 (2023) - [i200]Chien-yu Huang, Ke-Han Lu, Shih-Heng Wang, Chi-Yuan Hsiao, Chun-Yi Kuan, Haibin Wu, Siddhant Arora, Kai-Wei Chang, Jiatong Shi, Yifan Peng, Roshan S. Sharma, Shinji Watanabe, Bhiksha Ramakrishnan, Shady Shehata, Hung-yi Lee:
Dynamic-SUPERB: Towards A Dynamic, Collaborative, and Comprehensive Instruction-Tuning Benchmark for Speech. CoRR abs/2309.09510 (2023) - [i199]Yuan Tseng, Layne Berry, Yi-Ting Chen, I-Hsiang Chiu, Hsuan-Hao Lin, Max Liu, Puyuan Peng, Yi-Jen Shih, Hung-Yu Wang, Haibin Wu, Po-Yao Huang, Chun-Mao Lai, Shang-Wen Li, David Harwath, Yu Tsao, Shinji Watanabe, Abdelrahman Mohamed, Chi-Luen Feng, Hung-yi Lee:
AV-SUPERB: A Multi-Task Evaluation Benchmark for Audio-Visual Representation Models. CoRR abs/2309.10787 (2023) - [i198]Chun-Yi Kuan, Chen-An Li, Tsu-Yuan Hsu, Tse-Yang Lin, Ho-Lam Chung, Kai-Wei Chang, Shuo-Yiin Chang, Hung-yi Lee:
Towards General-Purpose Text-Instruction-Guided Voice Conversion. CoRR abs/2309.14324 (2023) - [i197]Dennis Y. Menn, Tzu-hsun Feng, Sriram Vishwanath, Hung-yi Lee:
Investigating Human-Identifiable Features Hidden in Adversarial Perturbations. CoRR abs/2309.16878 (2023) - [i196]Po-Chun Hsu, Ali Elkahky, Wei-Ning Hsu, Yossi Adi, Tu Anh Nguyen, Jade Copet, Emmanuel Dupoux, Hung-yi Lee, Abdelrahman Mohamed:
Low-Resource Self-Supervised Learning with SSL-Enhanced TTS. CoRR abs/2309.17020 (2023) - [i195]Kai-Wei Chang, Ming-Hsin Chen, Yun-Ping Lin, Jing Neng Hsu, Paul Kuo-Ming Huang, Chien-Yu Huang, Shang-Wen Li, Hung-yi Lee:
Prompting and Adapter Tuning for Self-supervised Encoder-Decoder Speech Model. CoRR abs/2310.02971 (2023) - [i194]Kuan-Po Huang, Chih-Kai Yang, Yu-Kuan Fu, Ewan Dunbar, Hung-yi Lee:
Zero Resource Code-switched Speech Benchmark Using Speech Utterance Pairs For Multiple Spoken Languages. CoRR abs/2310.03018 (2023) - [i193]Shih-Cheng Huang, Pin-Zu Li, Yu-Chi Hsu, Kuang-Ming Chen, Yu-Tung Lin, Shih-Kai Hsiao, Richard Tzong-Han Tsai, Hung-yi Lee:
Chat Vector: A Simple Approach to Equip LLMs With New Language Chat Capabilities. CoRR abs/2310.04799 (2023) - [i192]Jiatong Shi, William Chen, Dan Berrebbi, Hsiu-Hsuan Wang, Wei-Ping Huang, En-Pei Hu, Ho-Lam Chung, Xuankai Chang, Yuxun Tang, Shang-Wen Li, Abdelrahman Mohamed, Hung-yi Lee, Shinji Watanabe:
Findings of the 2023 ML-SUPERB Challenge: Pre-Training and Evaluation over More Languages and Beyond. CoRR abs/2310.05513 (2023) - [i191]David Cheng-Han Chiang, Hung-yi Lee:
A Closer Look into Automatic Evaluation Using Large Language Models. CoRR abs/2310.05657 (2023) - [i190]Hsuan Su, Cheng-Chu Cheng, Hua Farn, Shachi H. Kumar, Saurav Sahay, Shang-Tse Chen, Hung-yi Lee:
Learning from Red Teaming: Gender Bias Provocation and Mitigation in Large Language Models. CoRR abs/2310.11079 (2023) - [i189]Ming-Hao Hsu, Kai-Wei Chang, Shang-Wen Li, Hung-yi Lee:
An Exploration of In-Context Learning for Speech Language Model. CoRR abs/2310.12477 (2023) - [i188]Hsuan Su, Rebecca Qian, Chinnadhurai Sankar, Shahin Shayandeh, Shang-Tse Chen, Hung-yi Lee, Daniel M. Bikel:
Step by Step to Fairness: Attributing Societal Bias in Task-oriented Dialogue Systems. CoRR abs/2311.06513 (2023) - [i187]Haibin Wu, Heng-Cheng Kuo, Yu Tsao, Hung-yi Lee:
Scalable Ensemble-based Detection Method against Adversarial Attacks for speaker verification. CoRR abs/2312.08622 (2023) - [i186]Min-Han Shih, Ho-Lam Chung, Yu-Chi Pai, Ming-Hao Hsu, Guan-Ting Lin, Shang-Wen Li, Hung-Yi Lee:
GSQA: An End-to-End Model for Generative Spoken Question Answering. CoRR abs/2312.09781 (2023) - [i185]Fabian Ritter Gutierrez, Kuan-Po Huang, Dianwen Ng, Jeremy Heng Meng Wong, Hung-yi Lee, Eng Siong Chng, Nancy F. Chen:
Noise robust distillation of self-supervised speech models via correlation metrics. CoRR abs/2312.12153 (2023) - [i184]Guan-Ting Lin, Prashanth Gurunath Shivakumar, Ankur Gandhe, Chao-Han Huck Yang, Yile Gu, Shalini Ghosh, Andreas Stolcke, Hung-yi Lee, Ivan Bulyko:
Paralinguistics-Enhanced Large Language Modeling of Spoken Dialogue. CoRR abs/2312.15316 (2023) - 2022
- [j22]Hung-Yi Lee, Shinji Watanabe, Karen Livescu, Abdelrahman Mohamed, Tara N. Sainath:
Editorial Editorial of Special Issue on Self-Supervised Learning for Speech and Audio Processing. IEEE J. Sel. Top. Signal Process. 16(6): 1174-1178 (2022) - [j21]Abdelrahman Mohamed, Hung-yi Lee, Lasse Borgholt, Jakob D. Havtorn, Joakim Edin, Christian Igel, Katrin Kirchhoff, Shang-Wen Li, Karen Livescu, Lars Maaløe, Tara N. Sainath, Shinji Watanabe:
Self-Supervised Speech Representation Learning: A Review. IEEE J. Sel. Top. Signal Process. 16(6): 1179-1210 (2022) - [j20]Haibin Wu, Xu Li, Andy T. Liu, Zhiyong Wu, Helen Meng, Hung-Yi Lee:
Improving the Adversarial Robustness for Speaker Verification by Self-Supervised Learning. IEEE ACM Trans. Audio Speech Lang. Process. 30: 202-217 (2022) - [j19]Da-Rong Liu, Po-Chun Hsu, Yi-Chen Chen, Sung-Feng Huang, Shun-Po Chuang, Da-Yi Wu, Hung-yi Lee:
Learning Phone Recognition From Unpaired Audio and Phone Sequences Based on Generative Adversarial Network. IEEE ACM Trans. Audio Speech Lang. Process. 30: 230-243 (2022) - [j18]Sung-Feng Huang, Chyi-Jiunn Lin, Da-Rong Liu, Yi-Chen Chen, Hung-yi Lee:
Meta-TTS: Meta-Learning for Few-Shot Speaker Adaptive Text-to-Speech. IEEE ACM Trans. Audio Speech Lang. Process. 30: 1558-1571 (2022) - [j17]Yi-Long Liou, Jui-Yang Hsu, Chen-Sheng Chen, Alexander H. Liu, Hung-Yi Lee, Tsung-Te Liu:
A Fully Integrated 1.7mW Attention-Based Automatic Speech Recognition Processor. IEEE Trans. Circuits Syst. II Express Briefs 69(10): 4178-4182 (2022) - [c167]David Cheng-Han Chiang, Hung-Yi Lee:
On the Transferability of Pre-trained Language Models: A Study from Artificial Datasets. AAAI 2022: 10518-10525 - [c166]Chan-Jan Hsu, Hung-yi Lee, Yu Tsao:
XDBERT: Distilling Visual Information to BERT from Cross-Modal Systems to Improve Language Understanding. ACL (2) 2022: 479-489 - [c165]Hsiang-Sheng Tsai, Heng-Jui Chang, Wen-Chin Huang, Zili Huang, Kushal Lakhotia, Shu-Wen Yang, Shuyan Dong, Andy T. Liu, Cheng-I Lai, Jiatong Shi, Xuankai Chang, Phil Hall, Hsuan-Jui Chen, Shang-Wen Li, Shinji Watanabe, Abdelrahman Mohamed, Hung-yi Lee:
SUPERB-SG: Enhanced Speech processing Universal PERformance Benchmark for Semantic and Generative Capabilities. ACL (1) 2022: 8479-8492 - [c164]Haibin Wu, Po-Chun Hsu, Ji Gao, Shanshan Zhang, Shen Huang, Jian Kang, Zhiyong Wu, Helen Meng, Hung-Yi Lee:
Adversarial Sample Detection for Speaker Verification by Neural Vocoders. ICASSP 2022: 236-240 - [c163]Haibin Wu, Bo Zheng, Xu Li, Xixin Wu, Hung-Yi Lee, Helen Meng:
Characterizing the Adversarial Vulnerability of Speech self-Supervised Learning. ICASSP 2022: 3164-3168 - [c162]Yen Meng, Yi-Hui Chou, Andy T. Liu, Hung-yi Lee:
Don't Speak Too Fast: The Impact of Data Bias on Self-Supervised Speech Models. ICASSP 2022: 3258-3262 - [c161]Wen-Chin Huang, Shu-Wen Yang, Tomoki Hayashi, Hung-Yi Lee, Shinji Watanabe, Tomoki Toda:
S3PRL-VC: Open-Source Voice Conversion Framework with Self-Supervised Speech Representations. ICASSP 2022: 6552-6556 - [c160]Chien-yu Huang, Kai-Wei Chang, Hung-Yi Lee:
Toward Degradation-Robust Voice Conversion. ICASSP 2022: 6777-6781 - [c159]Heng-Jui Chang, Shu-Wen Yang, Hung-yi Lee:
Distilhubert: Speech Representation Learning by Layer-Wise Distillation of Hidden-Unit Bert. ICASSP 2022: 7087-7091 - [c158]Guan-Ting Lin, Chan-Jan Hsu, Da-Rong Liu, Hung-Yi Lee, Yu Tsao:
Analyzing The Robustness of Unsupervised Speech Recognition. ICASSP 2022: 8202-8206 - [c157]Haibin Wu, Heng-Cheng Kuo, Naijun Zheng, Kuo-Hsuan Hung, Hung-Yi Lee, Yu Tsao, Hsin-Min Wang, Helen Meng:
Partially Fake Audio Detection by Self-Attention-Based Fake Span Discovery. ICASSP 2022: 9236-9240 - [c156]Yang Zhang, Zhiqiang Lv, Haibin Wu, Shanshan Zhang, Pengfei Hu, Zhiyong Wu, Hung-yi Lee, Helen Meng:
MFA-Conformer: Multi-scale Feature Aggregation Conformer for Automatic Speaker Verification. INTERSPEECH 2022: 306-310 - [c155]Kuan-Po Huang, Yu-Kuan Fu, Yu Zhang, Hung-yi Lee:
Improving Distortion Robustness of Self-supervised Speech Processing Tasks with Domain Adaptation. INTERSPEECH 2022: 2193-2197 - [c154]Guan-Ting Lin, Shang-Wen Li, Hung-yi Lee:
Listen, Adapt, Better WER: Source-free Single-utterance Test-time Adaptation for Automatic Speech Recognition. INTERSPEECH 2022: 2198-2202 - [c153]Haibin Wu, Lingwei Meng, Jiawen Kang, Jinchao Li, Xu Li, Xixin Wu, Hung-yi Lee, Helen Meng:
Spoofing-Aware Speaker Verification by Multi-Level Fusion. INTERSPEECH 2022: 4357-4361 - [c152]Wei-Cheng Tseng, Wei-Tsung Kao, Hung-yi Lee:
DDOS: A MOS Prediction Framework utilizing Domain Adaptive Pre-training and Distribution of Opinion Scores. INTERSPEECH 2022: 4541-4545 - [c151]Wei-Ping Huang, Po-Chun Chen, Sung-Feng Huang, Hung-yi Lee:
Few Shot Cross-Lingual TTS Using Transferable Phoneme Embedding. INTERSPEECH 2022: 4566-4570 - [c150]Kai-Wei Chang, Wei-Cheng Tseng, Shang-Wen Li, Hung-yi Lee:
An Exploration of Prompt Tuning on Generative Spoken Language Model for Speech Processing Tasks. INTERSPEECH 2022: 5005-5009 - [c149]Wei-Cheng Tseng, Wei-Tsung Kao, Hung-yi Lee:
Membership Inference Attacks Against Self-supervised Speech Models. INTERSPEECH 2022: 5040-5044 - [c148]Guan-Ting Lin, Yung-Sung Chuang, Ho-Lam Chung, Shu-Wen Yang, Hsuan-Jui Chen, Shuyan Annie Dong, Shang-Wen Li, Abdelrahman Mohamed, Hung-yi Lee, Lin-Shan Lee:
DUAL: Discrete Spoken Unit Adaptive Learning for Textless Spoken Question Answering. INTERSPEECH 2022: 5165-5169 - [c147]Chih-Chiang Chang, Hung-yi Lee:
Exploring Continuous Integrate-and-Fire for Adaptive Simultaneous Speech Translation. INTERSPEECH 2022: 5175-5179 - [c146]Chih-Chiang Chang, Shun-Po Chuang, Hung-yi Lee:
Anticipation-Free Training for Simultaneous Machine Translation. IWSLT@ACL 2022: 43-61 - [c145]Hung-yi Lee, Shang-Wen Li, Thang Vu:
Meta Learning for Natural Language Processing: A Survey. NAACL-HLT 2022: 666-684 - [c144]Chin-Lun Fu, Zih-Ching Chen, Yun-Ru Lee, Hung-yi Lee:
AdapterBias: Parameter-efficient Token-dependent Representation Shift for Adapters in NLP Tasks. NAACL-HLT (Findings) 2022: 2608-2621 - [c143]Haibin Wu, Jiawen Kang, Lingwei Meng, Yang Zhang, Xixin Wu, Zhiyong Wu, Hung-yi Lee, Helen Meng:
Tackling Spoofing-Aware Speaker Verification with Multi-Model Fusion. Odyssey 2022: 92-99 - [c142]Wei-Tsung Kao, Yuan-Kuei Wu, Chia-Ping Chen, Zhi-Sheng Chen, Yu-Pao Tsai, Hung-Yi Lee:
On the Efficiency of Integrating Self-Supervised Learning and Meta-Learning for User-Defined Few-Shot Keyword Spotting. SLT 2022: 414-421 - [c141]Xuanjun Chen, Haibin Wu, Helen Meng, Hung-yi Lee, Jyh-Shing Roger Jang:
Push-Pull: Characterizing the Adversarial Robustness for Audio-Visual Active Speaker Detection. SLT 2022: 692-699 - [c140]Yi-Jen Shih, Hsuan-Fu Wang, Heng-Jui Chang, Layne Berry, Hung-yi Lee, David Harwath:
SpeechCLIP: Integrating Speech with Pre-Trained Vision and Language Model. SLT 2022: 715-722 - [c139]Tzu-hsun Feng, Shuyan Annie Dong, Ching-Feng Yeh, Shu-Wen Yang, Tzu-Quan Lin, Jiatong Shi, Kai-Wei Chang, Zili Huang, Haibin Wu, Xuankai Chang, Shinji Watanabe, Abdelrahman Mohamed, Shang-Wen Li, Hung-yi Lee:
Superb @ SLT 2022: Challenge on Generalization and Efficiency of Self-Supervised Speech Representation Learning. SLT 2022: 1096-1103 - [c138]Guan-Ting Lin, Chi-Luen Feng, Wei-Ping Huang, Yuan Tseng, Tzu-Han Lin, Chen-An Li, Hung-yi Lee, Nigel G. Ward:
On the Utility of Self-Supervised Models for Prosody-Related Tasks. SLT 2022: 1104-1111 - [c137]Kuan-Po Huang, Yu-Kuan Fu, Tsu-Yuan Hsu, Fabian Ritter Gutierrez, Fan-Lin Wang, Liang-Hsuan Tseng, Yu Zhang, Hung-yi Lee:
Improving Generalizability of Distilled Self-Supervised Speech Processing Models Under Distorted Settings. SLT 2022: 1112-1119 - [c136]Zih-Ching Chen, Chin-Lun Fu, Chih-Ying Liu, Shang-Wen (Daniel) Li, Hung-yi Lee:
Exploring Efficient-Tuning Methods in Self-Supervised Speech Models. SLT 2022: 1120-1127 - [c135]Yen Meng, Hsuan-Jui Chen, Jiatong Shi, Shinji Watanabe, Paola García, Hung-yi Lee, Hao Tang:
On Compressing Sequences for Self-Supervised Speech Models. SLT 2022: 1128-1135 - [e1]Kong Aik Lee, Hung-yi Lee, Yanfeng Lu, Minghui Dong:
13th International Symposium on Chinese Spoken Language Processing, ISCSLP 2022, Singapore, December 11-14, 2022. IEEE 2022, ISBN 979-8-3503-9796-3 [contents] - [i183]Chih-Chiang Chang, Shun-Po Chuang, Hung-yi Lee:
Anticipation-free Training for Simultaneous Translation. CoRR abs/2201.12868 (2022) - [i182]Haibin Wu, Heng-Cheng Kuo, Naijun Zheng, Kuo-Hsuan Hung, Hung-Yi Lee, Yu Tsao, Hsin-Min Wang, Helen Meng:
Partially Fake Audio Detection by Self-attention-based Fake Span Discovery. CoRR abs/2202.06684 (2022) - [i181]Guan-Ting Lin, Yung-Sung Chuang, Ho-Lam Chung, Shu-Wen Yang, Hsuan-Jui Chen, Shuyan Dong, Shang-Wen Li, Abdelrahman Mohamed, Hung-yi Lee, Lin-Shan Lee:
DUAL: Discrete Spoken Unit Adaptive Learning for Textless Spoken Question Answering. CoRR abs/2203.04911 (2022) - [i180]Kuan-Po Huang, Yuan-Kuei Wu, Hung-yi Lee:
Improving the transferability of speech separation by meta-learning. CoRR abs/2203.05882 (2022) - [i179]Hsiang-Sheng Tsai, Heng-Jui Chang, Wen-Chin Huang, Zili Huang, Kushal Lakhotia, Shu-Wen Yang, Shuyan Dong, Andy T. Liu, Cheng-I Jeff Lai, Jiatong Shi, Xuankai Chang, Phil Hall, Hsuan-Jui Chen, Shang-Wen Li, Shinji Watanabe, Abdelrahman Mohamed, Hung-yi Lee:
SUPERB-SG: Enhanced Speech processing Universal PERformance Benchmark for Semantic and Generative Capabilities. CoRR abs/2203.06849 (2022) - [i178]Guan-Ting Lin, Shang-Wen Li, Hung-yi Lee:
Listen, Adapt, Better WER: Source-free Single-utterance Test-time Adaptation for Automatic Speech Recognition. CoRR abs/2203.14222 (2022) - [i177]Yang Zhang, Zhiqiang Lv, Haibin Wu, Shanshan Zhang, Pengfei Hu, Zhiyong Wu, Hung-Yi Lee, Helen Meng:
MFA-Conformer: Multi-scale Feature Aggregation Conformer for Automatic Speaker Verification. CoRR abs/2203.15249 (2022) - [i176]Haibin Wu, Lingwei Meng, Jiawen Kang, Jinchao Li, Xu Li, Xixin Wu, Hung-yi Lee, Helen Meng:
Spoofing-Aware Speaker Verification by Multi-Level Fusion. CoRR abs/2203.15377 (2022) - [i175]Kuan-Po Huang, Yu-Kuan Fu, Yu Zhang, Hung-yi Lee:
Improving Distortion Robustness of Self-supervised Speech Processing Tasks with Domain Adaptation. CoRR abs/2203.16104 (2022) - [i174]Kai-Wei Chang, Wei-Cheng Tseng, Shang-Wen Li, Hung-yi Lee:
An Exploration of Prompt Tuning on Generative Spoken Language Model for Speech Processing Tasks. CoRR abs/2203.16773 (2022) - [i173]Fan-Lin Wang, Po-Chun Hsu, Da-Rong Liu, Hung-yi Lee:
Universal Adaptor: Converting Mel-Spectrograms Between Different Configurations for Speech Synthesis. CoRR abs/2204.00170 (2022) - [i172]Wei-Tsung Kao, Yuen-Kwei Wu, Chia-Ping Chen, Zhi-Sheng Chen, Yu-Pao Tsai, Hung-Yi Lee:
On the Efficiency of Integrating Self-supervised Learning and Meta-learning for User-defined Few-shot Keyword Spotting. CoRR abs/2204.00352 (2022) - [i171]Wei-Cheng Tseng, Wei-Tsung Kao, Hung-yi Lee:
DDOS: A MOS Prediction Framework utilizing Domain Adaptive Pre-training and Distribution of Opinion Scores. CoRR abs/2204.03219 (2022) - [i170]David Cheng-Han Chiang, Hung-Yi Lee:
Understanding, Detecting, and Separating Out-of-Distribution Samples and Adversarial Samples in Text Classification. CoRR abs/2204.04458 (2022) - [i169]David Cheng-Han Chiang, Hung-Yi Lee:
Re-Examining Human Annotations for Interpretable NLP. CoRR abs/2204.04580 (2022) - [i168]Chan-Jan Hsu, Hung-yi Lee, Yu Tsao:
XDBERT: Distilling Visual Information to BERT from Cross-Modal Systems to Improve Language Understanding. CoRR abs/2204.07316 (2022) - [i167]Chih-Chiang Chang, Hung-yi Lee:
Exploring Continuous Integrate-and-Fire for Adaptive Simultaneous Speech Translation. CoRR abs/2204.09595 (2022) - [i166]Po-Chun Hsu, Da-Rong Liu, Andy T. Liu, Hung-yi Lee:
Parallel Synthesis for Autoregressive Speech Generation. CoRR abs/2204.11806 (2022) - [i165]Chin-Lun Fu, Zih-Ching Chen, Yun-Ru Lee, Hung-yi Lee:
AdapterBias: Parameter-efficient Token-dependent Representation Shift for Adapters in NLP Tasks. CoRR abs/2205.00305 (2022) - [i164]Hung-yi Lee, Shang-Wen Li, Ngoc Thang Vu:
Meta Learning for Natural Language Processing: A Survey. CoRR abs/2205.01500 (2022) - [i163]Chi-Luen Feng, Po-Chun Hsu, Hung-yi Lee:
Silence is Sweeter Than Speech: Self-Supervised Model Using Silence to Store Speaker Information. CoRR abs/2205.03759 (2022) - [i162]Abdelrahman Mohamed, Hung-yi Lee, Lasse Borgholt, Jakob D. Havtorn, Joakim Edin, Christian Igel, Katrin Kirchhoff, Shang-Wen Li, Karen Livescu, Lars Maaløe, Tara N. Sainath, Shinji Watanabe:
Self-Supervised Speech Representation Learning: A Review. CoRR abs/2205.10643 (2022) - [i161]Chi-Liang Liu, Hung-yi Lee, Wen-tau Yih:
Structured Prompt Tuning. CoRR abs/2205.12309 (2022) - [i160]Dennis Y. Menn, Hung-yi Lee:
Searching for the Essence of Adversarial Perturbations. CoRR abs/2205.15357 (2022) - [i159]Hsuan Su, Po-Han Chi, Shih-Cheng Huang, Ho-Lam Chung, Saurav Sahay, Shang-Tse Chen, Hung-Yi Lee:
Few-shot Prompting Towards Controllable Response Generation. CoRR abs/2206.03931 (2022) - [i158]Haibin Wu, Jiawen Kang, Lingwei Meng, Yang Zhang, Xixin Wu, Zhiyong Wu, Hung-yi Lee, Helen Meng:
Tackling Spoofing-Aware Speaker Verification with Multi-Model Fusion. CoRR abs/2206.09131 (2022) - [i157]Wei-Ping Huang, Po-Chun Chen, Sung-Feng Huang, Hung-yi Lee:
Few-Shot Cross-Lingual TTS Using Transferable Phoneme Embedding. CoRR abs/2206.15427 (2022) - [i156]Da-Rong Liu, Po-Chun Hsu, Yi-Chen Chen, Sung-Feng Huang, Shun-Po Chuang, Da-Yi Wu, Hung-yi Lee:
Learning Phone Recognition from Unpaired Audio and Phone Sequences Based on Generative Adversarial Network. CoRR abs/2207.14568 (2022) - [i155]Tung-Yu Wu, Chen-An Li, Tzu-Han Lin, Tsu-Yuan Hsu, Hung-Yi Lee:
The Ability of Self-Supervised Speech Models for Audio Representations. CoRR abs/2209.12900 (2022) - [i154]Yi-Jen Shih, Hsuan-Fu Wang, Heng-Jui Chang, Layne Berry, Hung-yi Lee, David Harwath:
SpeechCLIP: Integrating Speech with Pre-Trained Vision and Language Model. CoRR abs/2210.00705 (2022) - [i153]Xuanjun Chen, Haibin Wu, Helen Meng, Hung-yi Lee, Jyh-Shing Roger Jang:
Push-Pull: Characterizing the Adversarial Robustness for Audio-Visual Active Speaker Detection. CoRR abs/2210.00753 (2022) - [i152]David Cheng-Han Chiang, Hung-yi Lee:
How Far Are We from Real Synonym Substitution Attacks? CoRR abs/2210.02844 (2022) - [i151]Zih-Ching Chen, Chin-Lun Fu, Chih-Ying Liu, Shang-Wen Li, Hung-yi Lee:
Exploring Efficient-tuning Methods in Self-supervised Speech Models. CoRR abs/2210.06175 (2022) - [i150]Guan-Ting Lin, Chi-Luen Feng, Wei-Ping Huang, Yuan Tseng, Tzu-Han Lin, Chen-An Li, Hung-yi Lee, Nigel G. Ward:
On the Utility of Self-supervised Models for Prosody-related Tasks. CoRR abs/2210.07185 (2022) - [i149]Yen Meng, Hsuan-Jui Chen, Jiatong Shi, Shinji Watanabe, Paola García, Hung-yi Lee, Hao Tang:
On Compressing Sequences for Self-Supervised Speech Models. CoRR abs/2210.07189 (2022) - [i148]Kuan-Po Huang, Yu-Kuan Fu, Tsu-Yuan Hsu, Fabian Ritter Gutierrez, Fan-Lin Wang, Liang-Hsuan Tseng, Yu Zhang, Hung-yi Lee:
Improving generalizability of distilled self-supervised speech processing models under distorted settings. CoRR abs/2210.07978 (2022) - [i147]Tzu-hsun Feng, Shuyan Annie Dong, Ching-Feng Yeh, Shu-Wen Yang, Tzu-Quan Lin, Jiatong Shi, Kai-Wei Chang, Zili Huang, Haibin Wu, Xuankai Chang, Shinji Watanabe, Abdelrahman Mohamed, Shang-Wen Li, Hung-yi Lee:
SUPERB @ SLT 2022: Challenge on Generalization and Efficiency of Self-Supervised Speech Representation Learning. CoRR abs/2210.08634 (2022) - [i146]Xuanjun Chen, Haibin Wu, Chung-Che Wang, Hung-yi Lee, Jyh-Shing Roger Jang:
Multimodal Transformer Distillation for Audio-Visual Synchronization. CoRR abs/2210.15563 (2022) - [i145]Chan-Jan Hsu, Ho-Lam Chung, Hung-yi Lee, Yu Tsao:
T5lephone: Bridging Speech and Text Self-supervised Models for Spoken Language Understanding via Phoneme level T5. CoRR abs/2211.00586 (2022) - [i144]Layne Berry, Yi-Jen Shih, Hsuan-Fu Wang, Heng-Jui Chang, Hung-yi Lee, David Harwath:
M-SpeechCLIP: Leveraging Large-Scale, Pre-Trained Models for Multilingual Speech to Image Retrieval. CoRR abs/2211.01180 (2022) - [i143]Hsuan-Jui Chen, Yen Meng, Hung-yi Lee:
Once-for-All Sequence Compression for Self-Supervised Speech Models. CoRR abs/2211.02332 (2022) - [i142]Jiatong Shi, Chan-Jan Hsu, Ho-Lam Chung, Dongji Gao, Paola García, Shinji Watanabe, Ann Lee, Hung-yi Lee:
Bridging Speech and Textual Pre-trained Models with Unsupervised ASR. CoRR abs/2211.03025 (2022) - [i141]Derek Xu, Shuyan Dong, Changhan Wang, Suyoun Kim, Zhaojiang Lin, Akshat Shrivastava, Shang-Wen Li, Liang-Hsuan Tseng, Alexei Baevski, Guan-Ting Lin, Hung-yi Lee, Yizhou Sun, Wei Wang:
Introducing Semantics into Speech Encoders. CoRR abs/2211.08402 (2022) - [i140]Tzu-Quan Lin, Hung-yi Lee, Hao Tang:
MelHuBERT: A simplified HuBERT on Mel spectrogram. CoRR abs/2211.09944 (2022) - [i139]Tzu-Quan Lin, Tsung-Huan Yang, Chun-Yao Chang, Kuang-Ming Chen, Tzu-hsun Feng, Hung-yi Lee, Hao Tang:
Compressing Transformer-based self-supervised models for speech processing. CoRR abs/2211.09949 (2022) - [i138]Tsu-Yuan Hsu, Chen-An Li, Tung-Yu Wu, Hung-yi Lee:
Model Extraction Attack against Self-supervised Speech Models. CoRR abs/2211.16044 (2022) - [i137]Dongji Gao, Jiatong Shi, Shun-Po Chuang, Leibny Paola García, Hung-yi Lee, Shinji Watanabe, Sanjeev Khudanpur:
EURO: ESPnet Unsupervised ASR Open-source Toolkit. CoRR abs/2211.17196 (2022) - [i136]Shih-Cheng Huang, Shih-Heng Wang, Min-Han Shih, Saurav Sahay, Hung-yi Lee:
General Framework for Self-Supervised Model Priming for Parameter-Efficient Fine-tuning. CoRR abs/2212.01032 (2022) - [i135]Zih-Ching Chen, Yu-Shun Sung, Hung-yi Lee:
CHAPTER: Exploiting Convolutional Neural Network Adapters for Self-supervised Speech Models. CoRR abs/2212.01282 (2022) - [i134]Suwon Shon, Siddhant Arora, Chyi-Jiunn Lin, Ankita Pasad, Felix Wu, Roshan S. Sharma, Wei-Lun Wu, Hung-Yi Lee, Karen Livescu, Shinji Watanabe:
SLUE Phase-2: A Benchmark Suite of Diverse Spoken Language Understanding Tasks. CoRR abs/2212.10525 (2022) - 2021
- [j16]Shun-Po Chuang, Alexander H. Liu, Tzu-Wei Sung, Hung-yi Lee:
Improving Automatic Speech Recognition and Speech Translation via Word Embedding Prediction. IEEE ACM Trans. Audio Speech Lang. Process. 29: 93-105 (2021) - [j15]Andy T. Liu, Shang-Wen Li, Hung-yi Lee:
TERA: Self-Supervised Learning of Transformer Encoder Representation for Speech. IEEE ACM Trans. Audio Speech Lang. Process. 29: 2351-2366 (2021) - [c134]Yu-Ching Chiu, Bo-Hao Chang, Tzu-Yu Chen, Cheng-Fu Yang, Nanyi Bi, Richard Tzong-Han Tsai, Hung-yi Lee, Jane Yung-jen Hsu:
Multi-modal User Intent Classification Under the Scenario of Smart Factory (Student Abstract). AAAI 2021: 15771-15772 - [c133]Shun-Po Chuang, Yung-Sung Chuang, Chih-Chiang Chang, Hung-yi Lee:
Investigating the Reordering Capability in CTC-based Non-Autoregressive End-to-End Speech Translation. ACL/IJCNLP (Findings) 2021: 1068-1077 - [c132]Xuankai Chang, Takashi Maekaku, Pengcheng Guo, Jing Shi, Yen-Ju Lu, Aswin Shanmugam Subramanian, Tianzi Wang, Shu-Wen Yang, Yu Tsao, Hung-yi Lee, Shinji Watanabe:
An Exploration of Self-Supervised Pretrained Representations for End-to-End Speech Recognition. ASRU 2021: 228-235 - [c131]Shun-Po Chuang, Heng-Jui Chang, Sung-Feng Huang, Hung-yi Lee:
Non-Autoregressive Mandarin-English Code-Switching Speech Recognition. ASRU 2021: 465-472 - [c130]Wei-Tsung Kao, Hung-yi Lee:
Is BERT a Cross-Disciplinary Knowledge Learner? A Surprising Finding of Pre-trained Models' Transferability. EMNLP (Findings) 2021: 2195-2208 - [c129]Tung-Yu Wu, Tsu-Yuan Hsu, Chen-An Li, Tzu-Han Lin, Hung-yi Lee:
The Efficacy of Self-Supervised Speech Models for Audio Representations. HEAR@NeurIPS 2021: 90-110 - [c128]Yuan-Kuei Wu, Kuan-Po Huang, Yu Tsao, Hung-yi Lee:
One Shot Learning for Speech Separation. ICASSP 2021: 5769-5773 - [c127]Yist Y. Lin, Chung-Ming Chien, Jheng-Hao Lin, Hung-yi Lee, Lin-Shan Lee:
Fragmentvc: Any-To-Any Voice Conversion by End-To-End Extracting and Fusing Fine-Grained Voice Fragments with Attention. ICASSP 2021: 5939-5943 - [c126]Yen-Hao Chen, Da-Yi Wu, Tsung-Han Wu, Hung-yi Lee:
Again-VC: A One-Shot Voice Conversion Using Activation Guidance and Adaptive Instance Normalization. ICASSP 2021: 5954-5958 - [c125]Haibin Wu, Xu Li, Andy T. Liu, Zhiyong Wu, Helen Meng, Hung-yi Lee:
Adversarial Defense for Automatic Speaker Verification by Cascaded Self-Supervised Learning Models. ICASSP 2021: 6718-6722 - [c124]Cheng-I Lai, Yung-Sung Chuang, Hung-Yi Lee, Shang-Wen Li, James R. Glass:
Semi-Supervised Spoken Language Understanding via Self-Supervised Speech and Language Model Pretraining. ICASSP 2021: 7468-7472 - [c123]Chung-Ming Chien, Jheng-Hao Lin, Chien-yu Huang, Po-Chun Hsu, Hung-yi Lee:
Investigating on Incorporating Pretrained and Learnable Speaker Representations for Multi-Speaker Multi-Style Text-to-Speech. ICASSP 2021: 8588-8592 - [c122]Jheng-Hao Lin, Yist Y. Lin, Chung-Ming Chien, Hung-yi Lee:
S2VC: A Framework for Any-to-Any Voice Conversion with Self-Supervised Pretrained Representations. Interspeech 2021: 836-840 - [c121]Shu-Wen Yang, Po-Han Chi, Yung-Sung Chuang, Cheng-I Jeff Lai, Kushal Lakhotia, Yist Y. Lin, Andy T. Liu, Jiatong Shi, Xuankai Chang, Guan-Ting Lin, Tzu-Hsien Huang, Wei-Cheng Tseng, Ko-tik Lee, Da-Rong Liu, Zili Huang, Shuyan Dong, Shang-Wen Li, Shinji Watanabe, Abdelrahman Mohamed, Hung-yi Lee:
SUPERB: Speech Processing Universal PERformance Benchmark. Interspeech 2021: 1194-1198 - [c120]Heng-Jui Chang, Hung-yi Lee, Lin-Shan Lee:
Towards Lifelong Learning of End-to-End ASR. Interspeech 2021: 2551-2555 - [c119]Wei-Cheng Tseng, Chien-yu Huang, Wei-Tsung Kao, Yist Y. Lin, Hung-yi Lee:
Utilizing Self-Supervised Representations for MOS Prediction. Interspeech 2021: 2781-2785 - [c118]Sung-Feng Huang, Shun-Po Chuang, Da-Rong Liu, Yi-Chen Chen, Gene-Ping Yang, Hung-yi Lee:
Stabilizing Label Assignment for Speech Separation by Self-Supervised Pre-Training. Interspeech 2021: 3056-3060 - [c117]Jingsong Wang, Yuxuan He, Chunyu Zhao, Qijie Shao, Wei-Wei Tu, Tom Ko, Hung-yi Lee, Lei Xie:
Auto-KWS 2021 Challenge: Task, Datasets, and Baselines. Interspeech 2021: 4244-4248 - [c116]Haibin Wu, Yang Zhang, Zhiyong Wu, Dong Wang, Hung-yi Lee:
Voting for the Right Answer: Adversarial Defense for Speaker Verification. Interspeech 2021: 4294-4298 - [c115]Hsuan Su, Jiun-Hao Jhan, Fan-Yun Sun, Saurav Sahay, Hung-yi Lee:
Put Chatbot into Its Interlocutor's Shoes: New Framework to Learn Chatbot Responding with Intention. NAACL-HLT 2021: 1559-1569 - [c114]Heng-Jui Chang, Alexander H. Liu, Hung-yi Lee, Lin-Shan Lee:
End-to-End Whispered Speech Recognition with Frequency-Weighted Approaches and Pseudo Whisper Pre-training. SLT 2021: 186-193 - [c113]Po-Han Chi, Pei-Hung Chung, Tsung-Han Wu, Chun-Cheng Hsieh, Yen-Hao Chen, Shang-Wen Li, Hung-yi Lee:
Audio Albert: A Lite Bert for Self-Supervised Learning of Audio Representation. SLT 2021: 344-350 - [c112]Chung-Ming Chien, Hung-yi Lee:
Hierarchical Prosody Modeling for Non-Autoregressive Speech Synthesis. SLT 2021: 446-453 - [c111]Tzu-hsien Huang, Jheng-Hao Lin, Hung-yi Lee:
How Far Are We from Robust Voice Conversion: A Survey. SLT 2021: 514-521 - [c110]Chien-yu Huang, Yist Y. Lin, Hung-yi Lee, Lin-Shan Lee:
Defending Your Voice: Adversarial Attack on Voice Conversion. SLT 2021: 552-559 - [i133]Haibin Wu, Xu Li, Andy T. Liu, Zhiyong Wu, Helen Meng, Hung-yi Lee:
Adversarial defense for automatic speaker verification by cascaded self-supervised learning models. CoRR abs/2102.07047 (2021) - [i132]Chung-Ming Chien, Jheng-Hao Lin, Chien-yu Huang, Po-Chun Hsu, Hung-yi Lee:
Investigating on Incorporating Pretrained and Learnable Speaker Representations for Multi-Speaker Multi-Style Text-to-Speech. CoRR abs/2103.04088 (2021) - [i131]Wei-Tsung Kao, Hung-Yi Lee:
Is BERT a Cross-Disciplinary Knowledge Learner? A Surprising Finding of Pre-trained Models' Transferability. CoRR abs/2103.07162 (2021) - [i130]Hsuan Su, Jiun-Hao Jhan, Fan-Yun Sun, Saurav Sahay, Hung-yi Lee:
Put Chatbot into Its Interlocutor's Shoes: New Framework to Learn Chatbot Responding with Intention. CoRR abs/2103.16429 (2021) - [i129]Jingsong Wang, Yuxuan He, Chunyu Zhao, Qijie Shao, Wei-Wei Tu, Tom Ko, Hung-yi Lee, Lei Xie:
Auto-KWS 2021 Challenge: Task, Datasets, and Baselines. CoRR abs/2104.00513 (2021) - [i128]Heng-Jui Chang, Hung-yi Lee, Lin-Shan Lee:
Towards Lifelong Learning of End-to-end ASR. CoRR abs/2104.01616 (2021) - [i127]Shun-Po Chuang, Heng-Jui Chang, Sung-Feng Huang, Hung-yi Lee:
Non-autoregressive Mandarin-English Code-switching Speech Recognition with Pinyin Mask-CTC and Word Embedding Regularization. CoRR abs/2104.02258 (2021) - [i126]Jheng-Hao Lin, Yist Y. Lin, Chung-Ming Chien, Hung-yi Lee:
S2VC: A Framework for Any-to-Any Voice Conversion with Self-Supervised Pretrained Representations. CoRR abs/2104.02901 (2021) - [i125]Wei-Cheng Tseng, Chien-yu Huang, Wei-Tsung Kao, Yist Y. Lin, Hung-yi Lee:
Utilizing Self-supervised Representations for MOS Prediction. CoRR abs/2104.03017 (2021) - [i124]Shu-Wen Yang, Po-Han Chi, Yung-Sung Chuang, Cheng-I Jeff Lai, Kushal Lakhotia, Yist Y. Lin, Andy T. Liu, Jiatong Shi, Xuankai Chang, Guan-Ting Lin, Tzu-Hsien Huang, Wei-Cheng Tseng, Ko-tik Lee, Da-Rong Liu, Zili Huang, Shuyan Dong, Shang-Wen Li, Shinji Watanabe, Abdelrahman Mohamed, Hung-yi Lee:
SUPERB: Speech processing Universal PERformance Benchmark. CoRR abs/2105.01051 (2021) - [i123]Yi-Chen Chen, Po-Han Chi, Shu-Wen Yang, Kai-Wei Chang, Jheng-Hao Lin, Sung-Feng Huang, Da-Rong Liu, Chi-Liang Liu, Cheng-Kuang Lee, Hung-yi Lee:
SpeechNet: A Universal Modularized Model for Speech Processing Tasks. CoRR abs/2105.03070 (2021) - [i122]Wei-Cheng Huang, Chien-yu Huang, Hung-yi Lee:
Improving Cross-Lingual Reading Comprehension with Self-Training. CoRR abs/2105.03627 (2021) - [i121]Shun-Po Chuang, Yung-Sung Chuang, Chih-Chiang Chang, Hung-yi Lee:
Investigating the Reordering Capability in CTC-based Non-Autoregressive End-to-End Speech Translation. CoRR abs/2105.04840 (2021) - [i120]Haibin Wu, Xu Li, Andy T. Liu, Zhiyong Wu, Helen Meng, Hung-yi Lee:
Improving the Adversarial Robustness for Speaker Verification by Self-Supervised Learning. CoRR abs/2106.00273 (2021) - [i119]Yung-Sung Chuang, Mingye Gao, Hongyin Luo, James R. Glass, Hung-Yi Lee, Yun-Nung Chen, Shang-Wen Li:
Mitigating Biases in Toxic Language Detection through Invariant Rationalization. CoRR abs/2106.07240 (2021) - [i118]Haibin Wu, Yang Zhang, Zhiyong Wu, Dong Wang, Hung-yi Lee:
Voting for the right answer: Adversarial defense for speaker verification. CoRR abs/2106.07868 (2021) - [i117]Kuan-Po Huang, Yuan-Kuei Wu, Hung-yi Lee:
Multi-accent Speech Separation with One Shot Learning. CoRR abs/2106.11713 (2021) - [i116]Haibin Wu, Po-Chun Hsu, Ji Gao, Shanshan Zhang, Shen Huang, Jian Kang, Zhiyong Wu, Helen Meng, Hung-yi Lee:
Spotting adversarial samples for speaker verification by neural vocoders. CoRR abs/2107.00309 (2021) - [i115]Zih-Yun Chiu, Yi-Lin Tuan, Hung-yi Lee, Li-Chen Fu:
Parallelized Reverse Curriculum Generation. CoRR abs/2108.02128 (2021) - [i114]David Cheng-Han Chiang, Hung-yi Lee:
On the Transferability of Pre-trained Language Models: A Study from Artificial Datasets. CoRR abs/2109.03537 (2021) - [i113]Heng-Jui Chang, Shu-Wen Yang, Hung-yi Lee:
DistilHuBERT: Speech Representation Learning by Layer-wise Distillation of Hidden-unit BERT. CoRR abs/2110.01900 (2021) - [i112]Liang-Hsuan Tseng, Yu-Kuan Fu, Heng-Jui Chang, Hung-yi Lee:
Mandarin-English Code-switching Speech Recognition with Self-supervised Speech Representation Models. CoRR abs/2110.03504 (2021) - [i111]Jiun-Hao Jhan, Chao-Peng Liu, Shyh-Kang Jeng, Hung-Yi Lee:
CheerBots: Chatbots toward Empathy and Emotionusing Reinforcement Learning. CoRR abs/2110.03949 (2021) - [i110]Xuankai Chang, Takashi Maekaku, Pengcheng Guo, Jing Shi, Yen-Ju Lu, Aswin Shanmugam Subramanian, Tianzi Wang, Shu-Wen Yang, Yu Tsao, Hung-yi Lee, Shinji Watanabe:
An Exploration of Self-Supervised Pretrained Representations for End-to-End Speech Recognition. CoRR abs/2110.04590 (2021) - [i109]Wen-Chin Huang, Shu-Wen Yang, Tomoki Hayashi, Hung-Yi Lee, Shinji Watanabe, Tomoki Toda:
S3PRL-VC: Open-source Voice Conversion Framework with Self-supervised Speech Representations. CoRR abs/2110.06280 (2021) - [i108]Chien-yu Huang, Kai-Wei Chang, Hung-yi Lee:
Toward Degradation-Robust Voice Conversion. CoRR abs/2110.07537 (2021) - [i107]Yen Meng, Yi-Hui Chou, Andy T. Liu, Hung-yi Lee:
Don't speak too fast: The impact of data bias on self-supervised speech models. CoRR abs/2110.07957 (2021) - [i106]Yi-Chen Chen, Shu-Wen Yang, Cheng-Kuang Lee, Simon See, Hung-yi Lee:
Speech Representation Learning Through Self-supervised Pretraining And Multi-task Finetuning. CoRR abs/2110.09930 (2021) - [i105]Sung-Feng Huang, Chyi-Jiunn Lin, Hung-yi Lee:
Meta-TTS: Meta-Learning for Few-Shot Speaker Adaptive Text-to-Speech. CoRR abs/2111.04040 (2021) - [i104]Haibin Wu, Bo Zheng, Xu Li, Xixin Wu, Hung-yi Lee, Helen Meng:
Characterizing the adversarial vulnerability of speech self-supervised learning. CoRR abs/2111.04330 (2021) - [i103]Wei-Cheng Tseng, Wei-Tsung Kao, Hung-yi Lee:
Membership Inference Attacks Against Self-supervised Speech Models. CoRR abs/2111.05113 (2021) - 2020
- [j14]Zixing Zhang, Dimitris N. Metaxas, Hung-yi Lee, Björn W. Schuller:
Guest Editorial Special Issue on Adversarial Learning in Computational Intelligence. IEEE Trans. Emerg. Top. Comput. Intell. 4(4): 414-416 (2020) - [c109]Che-Ping Tsai, Hung-yi Lee:
Order-Free Learning Alleviating Exposure Bias in Multi-Label Classification. AAAI 2020: 6038-6045 - [c108]Shun-Po Chuang, Tzu-Wei Sung, Alexander H. Liu, Hung-yi Lee:
Worse WER, but Better BLEU? Leveraging Word Embedding as Intermediate in Multitask End-to-End Speech Translation. ACL 2020: 5998-6003 - [c107]David Cheng-Han Chiang, Sung-Feng Huang, Hung-yi Lee:
Pretrained Language Model Embryology: The Birth of ALBERT. EMNLP (1) 2020: 6813-6828 - [c106]Chun-Hao Chao, Pin-Lun Hsu, Hung-Yi Lee, Yu-Chiang Frank Wang:
Self-Supervised Deep Learning for Fisheye Image Rectification. ICASSP 2020: 2248-2252 - [c105]Gene-Ping Yang, Szu-Lin Wu, Yao-Wen Mao, Hung-yi Lee, Lin-Shan Lee:
Interrupted and Cascaded Permutation Invariant Training for Speech Separation. ICASSP 2020: 6369-6373 - [c104]Andy T. Liu, Shu-Wen Yang, Po-Han Chi, Po-Chun Hsu, Hung-yi Lee:
Mockingjay: Unsupervised Speech Representation Learning with Deep Bidirectional Transformer Encoders. ICASSP 2020: 6419-6423 - [c103]Chung-Yi Li, Pei-Chieh Yuan, Hung-yi Lee:
What Does a Network Layer Hear? Analyzing Hidden Representations of End-to-End ASR Through Speech Synthesis. ICASSP 2020: 6434-6438 - [c102]Haibin Wu, Songxiang Liu, Helen Meng, Hung-yi Lee:
Defense Against Adversarial Attacks on Spoofing Countermeasures of ASV. ICASSP 2020: 6564-6568 - [c101]Alexander H. Liu, Tao Tu, Hung-yi Lee, Lin-Shan Lee:
Towards Unsupervised Speech Recognition and Synthesis with Quantized Speech Representation Learning. ICASSP 2020: 7259-7263 - [c100]Da-Yi Wu, Hung-Yi Lee:
One-Shot Voice Conversion by Vector Quantization. ICASSP 2020: 7734-7738 - [c99]Jui-Yang Hsu, Yuan-Jui Chen, Hung-yi Lee:
Meta Learning for End-To-End Low-Resource Speech Recognition. ICASSP 2020: 7844-7848 - [c98]Alexander H. Liu, Tzu-Wei Sung, Shun-Po Chuang, Hung-yi Lee, Lin-Shan Lee:
Sequence-to-Sequence Automatic Speech Recognition with Word Embedding Regularization and Fused Decoding. ICASSP 2020: 7879-7883 - [c97]Shun-Po Chuang, Tzu-Wei Sung, Hung-yi Lee:
Training Code-Switching Language Model with Monolingual Data. ICASSP 2020: 7949-7953 - [c96]Fan-Keng Sun, Cheng-Hao Ho, Hung-Yi Lee:
LAMOL: LAnguage MOdeling for Lifelong Language Learning. ICLR 2020 - [c95]Po-Chun Hsu, Hung-yi Lee:
WG-WaveNet: Real-Time High-Fidelity Speech Synthesis Without GPU. INTERSPEECH 2020: 210-214 - [c94]Yi-Chen Chen, Jui-Yang Hsu, Cheng-Kuang Lee, Hung-yi Lee:
DARTS-ASR: Differentiable Architecture Search for Multilingual Speech Recognition and Adaptation. INTERSPEECH 2020: 1803-1807 - [c93]Tao Tu, Yuan-Jui Chen, Alexander H. Liu, Hung-yi Lee:
Semi-Supervised Learning for Multi-Speaker Text-to-Speech Synthesis Using Discrete Speech Representation. INTERSPEECH 2020: 3191-3195 - [c92]Haibin Wu, Andy T. Liu, Hung-yi Lee:
Defense for Black-Box Attacks on Anti-Spoofing Models by Self-Supervised Learning. INTERSPEECH 2020: 3780-3784 - [c91]Shu-Wen Yang, Andy T. Liu, Hung-yi Lee:
Understanding Self-Attention of Self-Supervised Audio Transformers. INTERSPEECH 2020: 3785-3789 - [c90]Yung-Sung Chuang, Chi-Liang Liu, Hung-yi Lee, Lin-Shan Lee:
SpeechBERT: An Audio-and-Text Jointly Learned Language Model for End-to-End Spoken Question Answering. INTERSPEECH 2020: 4168-4172 - [c89]Da-Yi Wu, Yen-Hao Chen, Hung-yi Lee:
VQVC+: One-Shot Voice Conversion by Vector Quantization and U-Net Architecture. INTERSPEECH 2020: 4691-4695 - [c88]Chun-Hsing Lin, Siang-Ruei Wu, Hung-yi Lee, Yun-Nung Chen:
TaylorGAN: Neighbor-Augmented Policy Update Towards Sample-Efficient Natural Language Generation. NeurIPS 2020 - [i102]Wei-Tsung Kao, Tsung-Han Wu, Po-Han Chi, Chun-Cheng Hsieh, Hung-yi Lee:
Further Boosting BERT-based Models by Duplicating Existing Layers: Some Intriguing Phenomena inside BERT. CoRR abs/2001.09309 (2020) - [i101]Haibin Wu, Songxiang Liu, Helen Meng, Hung-yi Lee:
Defense against adversarial attacks on spoofing countermeasures of ASV. CoRR abs/2003.03065 (2020) - [i100]Chi-Liang Liu, Tsung-Yuan Hsu, Yung-Sung Chuang, Hung-yi Lee:
A Study of Cross-Lingual Ability and Language-specific Information in Multilingual BERT. CoRR abs/2004.09205 (2020) - [i99]Yau-Shian Wang, Hung-yi Lee, Yun-Nung Chen:
Learning Interpretable and Discrete Representations with Adversarial Training for Unsupervised Text Classification. CoRR abs/2004.13255 (2020) - [i98]Heng-Jui Chang, Alexander H. Liu, Hung-yi Lee, Lin-Shan Lee:
End-to-end Whispered Speech Recognition with Frequency-weighted Approaches and Layer-wise Transfer Learning. CoRR abs/2005.01972 (2020) - [i97]Yi-Chen Chen, Jui-Yang Hsu, Cheng-Kuang Lee, Hung-yi Lee:
DARTS-ASR: Differentiable Architecture Search for Multilingual Speech Recognition and Adaptation. CoRR abs/2005.07029 (2020) - [i96]Po-Chun Hsu, Hung-yi Lee:
WG-WaveNet: Real-Time High-Fidelity Speech Synthesis without GPU. CoRR abs/2005.07412 (2020) - [i95]Tao Tu, Yuan-Jui Chen, Alexander H. Liu, Hung-yi Lee:
Semi-supervised Learning for Multi-speaker Text-to-speech Synthesis Using Discrete Speech Representation. CoRR abs/2005.08024 (2020) - [i94]Po-Han Chi, Pei-Hung Chung, Tsung-Han Wu, Chun-Cheng Hsieh, Shang-wen Li, Hung-yi Lee:
Audio ALBERT: A Lite BERT for Self-supervised Learning of Audio Representation. CoRR abs/2005.08575 (2020) - [i93]Chien-yu Huang, Yist Y. Lin, Hung-yi Lee, Lin-Shan Lee:
Defending Your Voice: Adversarial Attack on Voice Conversion. CoRR abs/2005.08781 (2020) - [i92]Yuan-Kuei Wu, Chao-I Tuan, Hung-yi Lee, Yu Tsao:
SADDEL: Joint Speech Separation and Denoising Model based on Multitask Learning. CoRR abs/2005.09966 (2020) - [i91]Shun-Po Chuang, Tzu-Wei Sung, Alexander H. Liu, Hung-yi Lee:
Worse WER, but Better BLEU? Leveraging Word Embedding as Intermediate in Multitask End-to-End Speech Translation. CoRR abs/2005.10678 (2020) - [i90]Haibin Wu, Andy T. Liu, Hung-yi Lee:
Defense for Black-box Attacks on Anti-spoofing Models by Self-Supervised Learning. CoRR abs/2006.03214 (2020) - [i89]Shu-Wen Yang, Andy T. Liu, Hung-yi Lee:
Understanding Self-Attention of Self-Supervised Audio Transformers. CoRR abs/2006.03265 (2020) - [i88]Da-Yi Wu, Yen-Hao Chen, Hung-Yi Lee:
VQVC+: One-Shot Voice Conversion by Vector Quantization and U-Net architecture. CoRR abs/2006.04154 (2020) - [i87]Tsung-Han Wu, Chun-Cheng Hsieh, Yen-Hao Chen, Po-Han Chi, Hung-yi Lee:
Hand-crafted Attention is All You Need? A Study of Attention on Self-supervised Audio Transformer. CoRR abs/2006.05174 (2020) - [i86]Andy T. Liu, Shang-wen Li, Hung-yi Lee:
TERA: Self-Supervised Learning of Transformer Encoder Representation for Speech. CoRR abs/2007.06028 (2020) - [i85]Hung-yi Lee, Cheng-Hao Ho, Chien-Fu Lin, Chiung-Chih Chang, Chih-Wei Lee, Yau-Shian Wang, Tsung-Yuan Hsu, Kuan-Yu Chen:
Investigation of Sentiment Controllable Chatbot. CoRR abs/2007.07196 (2020) - [i84]David Cheng-Han Chiang, Sung-Feng Huang, Hung-yi Lee:
Pretrained Language Model Embryology: The Birth of ALBERT. CoRR abs/2010.02480 (2020) - [i83]Chi-Liang Liu, Tsung-Yuan Hsu, Yung-Sung Chuang, Chung-Yi Li, Hung-yi Lee:
Language Representation in Multilingual BERT and its applications to improve Cross-lingual Generalization. CoRR abs/2010.10041 (2020) - [i82]Chi-Liang Liu, Tsung-Yuan Hsu, Yung-Sung Chuang, Hung-yi Lee:
What makes multilingual BERT multilingual? CoRR abs/2010.10938 (2020) - [i81]Chi-Liang Liu, Hung-yi Lee:
Unsupervised Deep Learning based Multiple Choices Question Answering: Start Learning from Basic Knowledge. CoRR abs/2010.11003 (2020) - [i80]