Liu Yang is an assistant professor in the Key Laboratory of Intelligent Information Processing of Chinese Academy of Sciences(CAS), ICT, CAS.

He got his PhD from Institute of Computing Technology Chinese Academy of Sciences under the supervision of Prof. Qing He, and co-supervised by Prof. Xiang Ao. Here is his Google Scholar and DBLP Page.

From Feb 2022 to Feb 2023, he was a visiting scholar in the NExT Research Centre, National University of Singapore(NUS), adviced by Prof. Chua Tat-Seng. Also, he worked with Prof. Fuli Feng and Dr. Yunshan Ma.

Previously, he received the B.S. degree in Mathematics from Nanjing University (NJU) in 2017.

My research interest includes speech synthesis, neural machine translation and automatic music generation. I have published 20+ papers at the top international AI conferences such as WWW, ICLR, KDD.

If you like the template of this homepage, welcome to star and fork my open-sourced template version AcadHomepage .

🔥 News

  • 2024.03: 🎉 Two papers are accepted by ICLR 2024
  • 2023.05: 🎉 Five papers are accepted by ACL 2023
  • 2023.01: DiffSinger was introduced in a very popular video (2000k+ views) in Bilibili!
  • 2023.01: I join TikTok as a speech research scientist in Singapore!
  • 2022.02: I release a modern and responsive academic personal homepage template. Welcome to STAR and FORK!

📝 Publications

🎙 Speech Synthesis

NeurIPS 2019
sym

FastSpeech: Fast, Robust and Controllable Text to Speech
Yi Ren, Yangjun Ruan, Xu Tan, Tao Qin, Sheng Zhao, Zhou Zhao, Tie-Yan Liu

Project

  • FastSpeech is the first fully parallel end-to-end speech synthesis model.
  • Academic Impact: This work is included by many famous speech synthesis open-source projects, such as ESPNet . Our work are promoted by more than 20 media and forums, such as 机器之心InfoQ.
  • Industry Impact: FastSpeech has been deployed in Microsoft Azure TTS service and supports 49 more languages with state-of-the-art AI quality. It was also shown as a text-to-speech system acceleration example in NVIDIA GTC2020.
ICLR 2021
sym

FastSpeech 2: Fast and High-Quality End-to-End Text to Speech
Yi Ren, Chenxu Hu, Xu Tan, Tao Qin, Sheng Zhao, Zhou Zhao, Tie-Yan Liu

Project

ICLR 2024
sym

Mega-TTS 2: Boosting Prompting Mechanisms for Zero-Shot Speech Synthesis \ Ziyue Jiang, Jinglin Liu, Yi Ren, et al.

Project

  • This work has been deployed on many TikTok products.
  • Advandced zero-shot voice cloning model.
AAAI 2022
sym

DiffSinger: Singing Voice Synthesis via Shallow Diffusion Mechanism
Jinglin Liu, Chengxi Li, Yi Ren, Feiyang Chen, Zhou Zhao

NeurIPS 2021
sym

👄 TalkingFace & Avatar

ICLR 2024
sym

Real3D-Portrait: One-shot Realistic 3D Talking Portrait Synthesis, Zhenhui Ye, Tianyun Zhong, Yi Ren, et al. (Spotlight) Project | Code

📚 Machine Translation

🎼 Music & Dance Generation

🧑‍🎨 Generative Model

Others

🎖 Honors and Awards

📖 Educations

  • 2019.06 - 2022.04, Master, Zhejiang University, Hangzhou.
  • 2015.09 - 2019.06, Undergraduate, Chu Kochen Honors College, Zhejiang Univeristy, Hangzhou.
  • 2012.09 - 2015.06, Luqiao Middle School, Taizhou.

💬 Invited Talks

  • 2022.02, Hosted MLNLP seminar | [Video]
  • 2021.06, Audio & Speech Synthesis, Huawei internal talk
  • 2021.03, Non-autoregressive Speech Synthesis, PaperWeekly & biendata | [video]
  • 2020.12, Non-autoregressive Speech Synthesis, Huawei Noah’s Ark Lab internal talk

💻 Internships