Zhongqiu Wang

Orcid: 0000-0002-4204-9430

Affiliations:
  • Southern University of Science and Technology, Department of Computer Science and Engineering, Shenzhen, China
  • Carnegie Mellon University, Language Technologies Institute, Pittsburgh, PA, USA (2021 - 2024)
  • Google Research, Cambridge, MA, USA
  • Mitsubishi Electric Research Laboratories, Cambridge, MA, USA
  • Ohio State University, Department of Computer Science and Engineering, Columbus, OH, USA (PhD 2020)


According to our database1, Zhongqiu Wang authored at least 68 papers between 2015 and 2024.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

Online presence:

On csauthors.net:

Bibliography

2024
USDnet: Unsupervised Speech Dereverberation via Neural Forward Filtering.
IEEE ACM Trans. Audio Speech Lang. Process., 2024

Mixture to Mixture: Leveraging Close-Talk Mixtures as Weak-Supervision for Speech Separation.
IEEE Signal Process. Lett., 2024

Cross-Talk Reduction.
Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence, 2024

The Multimodal Information Based Speech Processing (MISP) 2023 Challenge: Audio-Visual Target Speaker Extraction.
Proceedings of the IEEE International Conference on Acoustics, 2024

Boosting Unknown-Number Speaker Separation with Transformer Decoder-Based Attractor.
Proceedings of the IEEE International Conference on Acoustics, 2024

Summary on the Multimodal Information-Based Speech Processing (MISP) 2023 Challenge.
Proceedings of the IEEE International Conference on Acoustics, 2024

2023
Software Design and User Interface of ESPnet-SE++: Speech Enhancement for Robust Speech Processing.
J. Open Source Softw., November, 2023

Software Design and User Interface of ESPnet-SE++: Speech Enhancement for Robust Speech Processing (espnet-v.202310).
Dataset, October, 2023

STFT-Domain Neural Speech Enhancement With Very Low Algorithmic Latency.
IEEE ACM Trans. Audio Speech Lang. Process., 2023

TF-GridNet: Integrating Full- and Sub-Band Modeling for Speech Separation.
IEEE ACM Trans. Audio Speech Lang. Process., 2023

Tackling the Cocktail Fork Problem for Separation and Transcription of Real-World Soundtracks.
IEEE ACM Trans. Audio Speech Lang. Process., 2023

The Multimodal Information Based Speech Processing (MISP) 2023 Challenge: Audio-Visual Target Speaker Extraction.
CoRR, 2023

The CHiME-7 DASR Challenge: Distant Meeting Transcription with Multiple Devices in Diverse Scenarios.
CoRR, 2023

Neural Speech Enhancement with Very Low Algorithmic Latency and Complexity via Integrated Full- and Sub-Band Modeling.
CoRR, 2023

Multi-Channel Target Speaker Extraction with Refinement: The WavLab Submission to the Second Clarity Enhancement Challenge.
CoRR, 2023

Exploring the Integration of Speech Separation and Recognition with Self-Supervised Learning Representation.
Proceedings of the IEEE Workshop on Applications of Signal Processing to Audio and Acoustics, 2023

UNSSOR: Unsupervised Neural Speech Separation by Leveraging Over-determined Training Mixtures.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023

FNeural Speech Enhancement with Very Low Algorithmic Latency and Complexity via Integrated full- and sub-band Modeling.
Proceedings of the IEEE International Conference on Acoustics, 2023

TF-GRIDNET: Making Time-Frequency Domain Models Great Again for Monaural Speaker Separation.
Proceedings of the IEEE International Conference on Acoustics, 2023

Multi-Channel Speaker Extraction with Adversarial Training: The Wavlab Submission to The Clarity ICASSP 2023 Grand Challenge.
Proceedings of the IEEE International Conference on Acoustics, 2023

Toward Universal Speech Enhancement For Diverse Input Conditions.
Proceedings of the IEEE Automatic Speech Recognition and Understanding Workshop, 2023

A Single Speech Enhancement Model Unifying Dereverberation, Denoising, Speaker Counting, Separation, And Extraction.
Proceedings of the IEEE Automatic Speech Recognition and Understanding Workshop, 2023

2022
Neural Spectrospatial Filtering.
IEEE ACM Trans. Audio Speech Lang. Process., 2022

Improving Frame-Online Neural Speech Enhancement With Overlapped-Frame Prediction.
IEEE Signal Process. Lett., 2022

ESPnet-SE++: Speech Enhancement for Robust Speech Recognition, Translation, and Understanding.
Proceedings of the 23rd Annual Conference of the International Speech Communication Association, 2022

Localization based Sequential Grouping for Continuous Speech Separation.
Proceedings of the IEEE International Conference on Acoustics, 2022

Locate This, Not that: Class-Conditioned Sound Event DOA Estimation.
Proceedings of the IEEE International Conference on Acoustics, 2022

The Cocktail Fork Problem: Three-Stem Audio Separation for Real-World Soundtracks.
Proceedings of the IEEE International Conference on Acoustics, 2022

Conditional Diffusion Probabilistic Model for Speech Enhancement.
Proceedings of the IEEE International Conference on Acoustics, 2022

Towards Low-Distortion Multi-Channel Speech Enhancement: The ESPNET-Se Submission to the L3DAS22 Challenge.
Proceedings of the IEEE International Conference on Acoustics, 2022

2021
Multi-microphone Complex Spectral Mapping for Utterance-wise and Continuous Speech Separation.
IEEE ACM Trans. Audio Speech Lang. Process., 2021

Convolutive Prediction for Monaural Speech Dereverberation and Noisy-Reverberant Speaker Separation.
IEEE ACM Trans. Audio Speech Lang. Process., 2021

On the Compensation Between Magnitude and Phase in Speech Separation.
IEEE Signal Process. Lett., 2021

Leveraging Low-Distortion Target Estimates for Improved Speech Enhancement.
CoRR, 2021

Anomalous Sound Detection Using Attentive Neural Processes.
Proceedings of the IEEE Workshop on Applications of Signal Processing to Audio and Acoustics, 2021

Convolutive Prediction for Reverberant Speech Separation.
Proceedings of the IEEE Workshop on Applications of Signal Processing to Audio and Acoustics, 2021

Sequential Multi-Frame Neural Beamforming for Speech Separation and Enhancement.
Proceedings of the IEEE Spoken Language Technology Workshop, 2021

Count And Separate: Incorporating Speaker Counting For Continuous Speaker Separation.
Proceedings of the IEEE International Conference on Acoustics, 2021

2020
Complex Spectral Mapping for Single- and Multi-Channel Speech Enhancement and Robust ASR.
IEEE ACM Trans. Audio Speech Lang. Process., 2020

Deep Learning Based Target Cancellation for Speech Dereverberation.
IEEE ACM Trans. Audio Speech Lang. Process., 2020

Robust Speaker Recognition Based on Single-Channel and Multi-Channel Speech Enhancement.
IEEE ACM Trans. Audio Speech Lang. Process., 2020

Multi-microphone Complex Spectral Mapping for Utterance-wise and Continuous Speaker Separation.
CoRR, 2020

Multi-Microphone Complex Spectral Mapping for Speech Dereverberation.
Proceedings of the 2020 IEEE International Conference on Acoustics, 2020

2019
Two-Stage Deep Learning for Noisy-Reverberant Speech Enhancement.
IEEE ACM Trans. Audio Speech Lang. Process., 2019

Robust Speaker Localization Guided by Deep Learning-Based Time-Frequency Masking.
IEEE ACM Trans. Audio Speech Lang. Process., 2019

Combining Spectral and Spatial Features for Deep Learning Based Blind Speaker Separation.
IEEE ACM Trans. Audio Speech Lang. Process., 2019

Alternating Between Spectral and Spatial Estimation for Speech Separation and Enhancement.
CoRR, 2019

Deep Learning Based Multi-Channel Speaker Recognition in Noisy and Reverberant Environments.
Proceedings of the 20th Annual Conference of the International Speech Communication Association, 2019

Deep Learning Based Phase Reconstruction for Speaker Separation: A Trigonometric Perspective.
Proceedings of the IEEE International Conference on Acoustics, 2019

2018
Robust TDOA Estimation Based on Time-Frequency Masking and Deep Neural Networks.
Proceedings of the 19th Annual Conference of the International Speech Communication Association, 2018

All-Neural Multi-Channel Speech Enhancement.
Proceedings of the 19th Annual Conference of the International Speech Communication Association, 2018

Integrating Spectral and Spatial Features for Multi-Channel Speaker Separation.
Proceedings of the 19th Annual Conference of the International Speech Communication Association, 2018

End-to-End Speech Separation with Unfolded Iterative Phase Reconstruction.
Proceedings of the 19th Annual Conference of the International Speech Communication Association, 2018

On Spatial Features for Supervised Speech Separation and its Application to Beamforming and Robust ASR.
Proceedings of the 2018 IEEE International Conference on Acoustics, 2018

Mask Weighted Stft Ratios for Relative Transfer Function Estimation and ITS Application to Robust ASR.
Proceedings of the 2018 IEEE International Conference on Acoustics, 2018

Alternative Objective Functions for Deep Clustering.
Proceedings of the 2018 IEEE International Conference on Acoustics, 2018

Multi-Channel Deep Clustering: Discriminative Spectral and Spatial Embeddings for Speaker-Independent Speech Separation.
Proceedings of the 2018 IEEE International Conference on Acoustics, 2018

2017
Speech emotion recognition based on Gaussian Mixture Models and Deep Neural Networks.
Proceedings of the 2017 Information Theory and Applications Workshop, 2017

A two-stage algorithm for noisy and reverberant speech enhancement.
Proceedings of the 2017 IEEE International Conference on Acoustics, 2017

A speech enhancement algorithm by iterating single- and multi-microphone processing and its application to robust ASR.
Proceedings of the 2017 IEEE International Conference on Acoustics, 2017

Unsupervised speaker adaptation of batch normalized acoustic models for robust ASR.
Proceedings of the 2017 IEEE International Conference on Acoustics, 2017

Recurrent deep stacking networks for supervised speech separation.
Proceedings of the 2017 IEEE International Conference on Acoustics, 2017

Learning utterance-level representations for speech emotion and age/gender recognition using deep neural networks.
Proceedings of the 2017 IEEE International Conference on Acoustics, 2017

2016
A Joint Training Framework for Robust Automatic Speech Recognition.
IEEE ACM Trans. Audio Speech Lang. Process., 2016

Phoneme-specific speech separation.
Proceedings of the 2016 IEEE International Conference on Acoustics, 2016

Robust speech recognition from ratio masks.
Proceedings of the 2016 IEEE International Conference on Acoustics, 2016

2015
Joint training of speech separation, filterbank and acoustic model for robust automatic speech recognition.
Proceedings of the 16th Annual Conference of the International Speech Communication Association, 2015

Combining spectral feature mapping and multi-channel model-based source separation for noise-robust automatic speech recognition.
Proceedings of the 2015 IEEE Workshop on Automatic Speech Recognition and Understanding, 2015


  Loading...