default search action
Marc Delcroix
Person information
Refine list
refinements active!
zoomed in on ?? of ?? records
view refined list in
export refined list as
2020 – today
- 2024
- [j31]Takanori Ashihara, Marc Delcroix, Yusuke Ijima, Makio Kashino:
Unveiling the Linguistic Capabilities of a Self-Supervised Speech Model Through Cross-Lingual Benchmark and Layer- Wise Similarity Analysis. IEEE Access 12: 98835-98855 (2024) - [j30]Tsubasa Ochiai, Kazuma Iwamoto, Marc Delcroix, Rintaro Ikeshita, Hiroshi Sato, Shoko Araki, Shigeru Katagiri:
Rethinking Processing Distortions: Disentangling the Impact of Speech Enhancement Errors on Speech Recognition Performance. IEEE ACM Trans. Audio Speech Lang. Process. 32: 3589-3602 (2024) - [c160]Rino Kimura, Tomohiro Nakatani, Naoyuki Kamo, Marc Delcroix, Shoko Araki, Tetsuya Ueda, Shoji Makino:
Diffusion Model-Based MIMO Speech Denoising and Dereverberation. ICASSP Workshops 2024: 455-459 - [c159]Junyi Peng, Marc Delcroix, Tsubasa Ochiai, Oldrich Plchot, Takanori Ashihara, Shoko Araki, Jan Cernocký:
Probing Self-Supervised Learning Models With Target Speech Extraction. ICASSP Workshops 2024: 535-539 - [c158]Keigo Wakayama, Tsubasa Ochiai, Marc Delcroix, Masahiro Yasuda, Shoichiro Saito, Shoko Araki, Akira Nakayama:
Online Target Sound Extraction with Knowledge Distillation from Partially Non-Causal Teacher. ICASSP 2024: 561-565 - [c157]Hao Shi, Naoyuki Kamo, Marc Delcroix, Tomohiro Nakatani, Shoko Araki:
Ensemble Inference for Diffusion Model-Based Speech Enhancement. ICASSP Workshops 2024: 735-739 - [c156]Thilo von Neumann, Christoph Böddeker, Tobias Cord-Landwehr, Marc Delcroix, Reinhold Haeb-Umbach:
Meeting Recognition with Continuous Speech Separation and Transcription-Supported Diarization. ICASSP Workshops 2024: 775-779 - [c155]Takanori Ashihara, Marc Delcroix, Takafumi Moriya, Kohei Matsuura, Taichi Asami, Yusuke Ijima:
What Do Self-Supervised Speech and Speaker Models Learn? New Findings from a Cross Model Layer-Wise Analysis. ICASSP 2024: 10166-10170 - [c154]Junyi Peng, Marc Delcroix, Tsubasa Ochiai, Oldrich Plchot, Shoko Araki, Jan Cernocký:
Target Speech Extraction with Pre-Trained Self-Supervised Learning Models. ICASSP 2024: 10421-10425 - [c153]Hanako Segawa, Tsubasa Ochiai, Marc Delcroix, Tomohiro Nakatani, Rintaro Ikeshita, Shoko Araki, Takeshi Yamada, Shoji Makino:
Neural Network-Based Virtual Microphone Estimation with Virtual Microphone and Beamformer-Level Multi-Task Loss. ICASSP 2024: 11021-11025 - [c152]Kazuma Iwamoto, Tsubasa Ochiai, Marc Delcroix, Rintaro Ikeshita, Hiroshi Sato, Shoko Araki, Shigeru Katagiri:
How Does End-To-End Speech Recognition Training Impact Speech Enhancement Artifacts? ICASSP 2024: 11031-11035 - [c151]Naohiro Tawara, Marc Delcroix, Atsushi Ando, Atsunori Ogawa:
NTT Speaker Diarization System for Chime-7: Multi-Domain, Multi-Microphone end-to-end and Vector Clustering Diarization. ICASSP 2024: 11281-11285 - [c150]Kenichi Fujita, Hiroshi Sato, Takanori Ashihara, Hiroki Kanagawa, Marc Delcroix, Takafumi Moriya, Yusuke Ijima:
Noise-Robust Zero-Shot Text-to-Speech Synthesis Conditioned on Self-Supervised Speech-Representation Model with Adapters. ICASSP 2024: 11471-11475 - [c149]Dominik Klement, Mireia Díez, Federico Landini, Lukás Burget, Anna Silnova, Marc Delcroix, Naohiro Tawara:
Discriminative Training of VBx Diarization. ICASSP 2024: 11871-11875 - [c148]William Chen, Takatomo Kano, Atsunori Ogawa, Marc Delcroix, Shinji Watanabe:
Train Long and Test Long: Leveraging Full Document Contexts in Speech Processing. ICASSP 2024: 13066-13070 - [c147]Tomohiro Nakatani, Naoyuki Kamo, Marc Delcroix, Shoko Araki:
Multi-Stream Diffusion Model for Probabilistic Integration of Model-Based and Data-Driven Speech Enhancement. IWAENC 2024: 65-69 - [c146]Carlos Hernandez-Olivan, Marc Delcroix, Tsubasa Ochiai, Naohiro Tawara, Tomohiro Nakatani, Shoko Araki:
Interaural Time Difference Loss for Binaural Target Sound Extraction. IWAENC 2024: 210-214 - [i73]Kenichi Fujita, Hiroshi Sato, Takanori Ashihara, Hiroki Kanagawa, Marc Delcroix, Takafumi Moriya, Yusuke Ijima:
Noise-robust zero-shot text-to-speech synthesis conditioned on self-supervised speech-representation model with adapters. CoRR abs/2401.05111 (2024) - [i72]Takanori Ashihara, Marc Delcroix, Takafumi Moriya, Kohei Matsuura, Taichi Asami, Yusuke Ijima:
What Do Self-Supervised Speech and Speaker Models Learn? New Findings From a Cross Model Layer-Wise Analysis. CoRR abs/2401.17632 (2024) - [i71]Marvin Tammen, Tsubasa Ochiai, Marc Delcroix, Tomohiro Nakatani, Shoko Araki, Simon Doclo:
Array Geometry-Robust Attention-Based Neural Beamformer for Moving Speakers. CoRR abs/2402.03058 (2024) - [i70]Junyi Peng, Marc Delcroix, Tsubasa Ochiai, Oldrich Plchot, Shoko Araki, Jan Cernocký:
Target Speech Extraction with Pre-trained Self-supervised Learning Models. CoRR abs/2402.13199 (2024) - [i69]Junyi Peng, Marc Delcroix, Tsubasa Ochiai, Oldrich Plchot, Takanori Ashihara, Shoko Araki, Jan Cernocký:
Probing Self-supervised Learning Models with Target Speech Extraction. CoRR abs/2402.13200 (2024) - [i68]Tsubasa Ochiai, Kazuma Iwamoto, Marc Delcroix, Rintaro Ikeshita, Hiroshi Sato, Shoko Araki, Shigeru Katagiri:
Rethinking Processing Distortions: Disentangling the Impact of Speech Enhancement Errors on Speech Recognition Performance. CoRR abs/2404.14860 (2024) - [i67]Atsunori Ogawa, Naoyuki Kamo, Kohei Matsuura, Takanori Ashihara, Takafumi Moriya, Takatomo Kano, Naohiro Tawara, Marc Delcroix:
Applying LLMs for Rescoring N-best ASR Hypotheses of Casual Conversations: Effects of Domain Adaptation and Context Carry-over. CoRR abs/2406.18972 (2024) - [i66]Kenichi Fujita, Takanori Ashihara, Marc Delcroix, Yusuke Ijima:
Lightweight Zero-shot Text-to-Speech with Mixture of Adapters. CoRR abs/2407.01291 (2024) - [i65]Hiroshi Sato, Takafumi Moriya, Masato Mimura, Shota Horiguchi, Tsubasa Ochiai, Takanori Ashihara, Atsushi Ando, Kentaro Shinayama, Marc Delcroix:
SpeakerBeam-SS: Real-time Target Speaker Extraction with Lightweight Conv-TasNet and State Space Modeling. CoRR abs/2407.01857 (2024) - [i64]Kohei Matsuura, Takanori Ashihara, Takafumi Moriya, Masato Mimura, Takatomo Kano, Atsunori Ogawa, Marc Delcroix:
Sentence-wise Speech Summarization: Task, Datasets, and End-to-End Modeling with LM Knowledge Distillation. CoRR abs/2408.00205 (2024) - [i63]Carlos Hernandez-Olivan, Marc Delcroix, Tsubasa Ochiai, Naohiro Tawara, Tomohiro Nakatani, Shoko Araki:
Interaural time difference loss for binaural target sound extraction. CoRR abs/2408.00344 (2024) - [i62]Shota Horiguchi, Atsushi Ando, Takafumi Moriya, Takanori Ashihara, Hiroshi Sato, Naohiro Tawara, Marc Delcroix:
Recursive Attentive Pooling for Extracting Speaker Embeddings from Multi-Speaker Recordings. CoRR abs/2408.17142 (2024) - [i61]Carlos Hernandez-Olivan, Marc Delcroix, Tsubasa Ochiai, Daisuke Niizumi, Naohiro Tawara, Tomohiro Nakatani, Shoko Araki:
SoundBeam meets M2D: Target Sound Extraction with Audio Foundation Model. CoRR abs/2409.12528 (2024) - [i60]Takafumi Moriya, Shota Horiguchi, Marc Delcroix, Ryo Masumura, Takanori Ashihara, Hiroshi Sato, Kohei Matsuura, Masato Mimura:
Alignment-Free Training for Transducer-based Multi-Talker ASR. CoRR abs/2409.20301 (2024) - [i59]Alexis Plaquet, Naohiro Tawara, Marc Delcroix, Shota Horiguchi, Atsushi Ando, Shoko Araki:
Mamba-based Segmentation Model for Speaker Diarization. CoRR abs/2410.06459 (2024) - [i58]Takanori Ashihara, Takafumi Moriya, Shota Horiguchi, Junyi Peng, Tsubasa Ochiai, Marc Delcroix, Kohei Matsuura, Hiroshi Sato:
Investigation of Speaker Representation for Target-Speaker Speech Processing. CoRR abs/2410.11243 (2024) - [i57]Shota Horiguchi, Takafumi Moriya, Atsushi Ando, Takanori Ashihara, Hiroshi Sato, Naohiro Tawara, Marc Delcroix:
Guided Speaker Embedding. CoRR abs/2410.12182 (2024) - 2023
- [j29]Takafumi Moriya, Hiroshi Sato, Tsubasa Ochiai, Marc Delcroix, Takahiro Shinozaki:
Streaming End-to-End Target-Speaker Automatic Speech Recognition and Activity Detection. IEEE Access 11: 13906-13917 (2023) - [j28]Katerina Zmolíková, Marc Delcroix, Tsubasa Ochiai, Keisuke Kinoshita, Jan Cernocký, Dong Yu:
Neural Target Speech Extraction: An overview. IEEE Signal Process. Mag. 40(3): 8-29 (2023) - [j27]Marc Delcroix, Jorge Bennasar Vázquez, Tsubasa Ochiai, Keisuke Kinoshita, Yasunori Ohishi, Shoko Araki:
SoundBeam: Target Sound Extraction Conditioned on Sound-Class Labels and Enrollment Clues for Increased Performance and Continuous Learning. IEEE ACM Trans. Audio Speech Lang. Process. 31: 121-136 (2023) - [j26]Thilo von Neumann, Keisuke Kinoshita, Christoph Böddeker, Marc Delcroix, Reinhold Haeb-Umbach:
Segment-Less Continuous Speech Separation of Meetings: Training and Evaluation Criteria. IEEE ACM Trans. Audio Speech Lang. Process. 31: 576-589 (2023) - [j25]Tsubasa Ochiai, Marc Delcroix, Tomohiro Nakatani, Shoko Araki:
Mask-Based Neural Beamforming for Moving Speakers With Self-Attention-Based Tracking. IEEE ACM Trans. Audio Speech Lang. Process. 31: 835-848 (2023) - [c145]Takatomo Kano, Atsunori Ogawa, Marc Delcroix, Kohei Matsuura, Takanori Ashihara, William Chen, Shinji Watanabe:
Summarize While Translating: Universal Model With Parallel Decoding for Summarization and Translation. ASRU 2023: 1-8 - [c144]Roshan S. Sharma, William Chen, Takatomo Kano, Ruchira Sharma, Siddhant Arora, Shinji Watanabe, Atsunori Ogawa, Marc Delcroix, Rita Singh, Bhiksha Raj:
Espnet-Summ: Introducing a Novel Large Dataset, Toolkit, and a Cross-Corpora Evaluation of Speech Summarization Systems. ASRU 2023: 1-8 - [c143]Takatomo Kano, Atsunori Ogawa, Marc Delcroix, Roshan S. Sharma, Kohei Matsuura, Shinji Watanabe:
Speech Summarization of Long Spoken Document: Improving Memory Efficiency of Speech/Text Encoders. ICASSP 2023: 1-5 - [c142]Kohei Matsuura, Takanori Ashihara, Takafumi Moriya, Tomohiro Tanaka, Atsunori Ogawa, Marc Delcroix, Ryo Masumura:
Leveraging Large Text Corpora For End-To-End Speech Summarization. ICASSP 2023: 1-5 - [c141]Thilo von Neumann, Christoph Böddeker, Keisuke Kinoshita, Marc Delcroix, Reinhold Haeb-Umbach:
On Word Error Rate Definitions and Their Efficient Computation for Multi-Speaker Speech Recognition Systems. ICASSP 2023: 1-5 - [c140]Atsunori Ogawa, Takafumi Moriya, Naoyuki Kamo, Naohiro Tawara, Marc Delcroix:
Iterative Shallow Fusion of Backward Language Model for End-To-End Speech Recognition. ICASSP 2023: 1-5 - [c139]Naoyuki Kamo, Marc Delcroix, Tomohiro Nakatani:
Target Speech Extraction with Conditional Diffusion Model. INTERSPEECH 2023: 176-180 - [c138]Hiroshi Sato, Ryo Masumura, Tsubasa Ochiai, Marc Delcroix, Takafumi Moriya, Takanori Ashihara, Kentaro Shinayama, Saki Mizuno, Mana Ihori, Tomohiro Tanaka, Nobukatsu Hojo:
Downstream Task Agnostic Speech Enhancement with Self-Supervised Representation Loss. INTERSPEECH 2023: 854-858 - [c137]Takafumi Moriya, Hiroshi Sato, Tsubasa Ochiai, Marc Delcroix, Takanori Ashihara, Kohei Matsuura, Tomohiro Tanaka, Ryo Masumura, Atsunori Ogawa, Taichi Asami:
Knowledge Distillation for Neural Transducer-based Target-Speaker ASR: Exploiting Parallel Mixture/Single-Talker Speech Data. INTERSPEECH 2023: 899-903 - [c136]Takanori Ashihara, Takafumi Moriya, Kohei Matsuura, Tomohiro Tanaka, Yusuke Ijima, Taichi Asami, Marc Delcroix, Yukinori Honma:
SpeechGLUE: How Well Can Self-Supervised Speech Models Capture Linguistic Knowledge? INTERSPEECH 2023: 2888-2892 - [c135]Kohei Matsuura, Takanori Ashihara, Takafumi Moriya, Tomohiro Tanaka, Takatomo Kano, Atsunori Ogawa, Marc Delcroix:
Transfer Learning from Pre-trained Language Models Improves End-to-End Speech Summarization. INTERSPEECH 2023: 2943-2947 - [c134]Marc Delcroix, Naohiro Tawara, Mireia Díez, Federico Landini, Anna Silnova, Atsunori Ogawa, Tomohiro Nakatani, Lukás Burget, Shoko Araki:
Multi-Stream Extension of Variational Bayesian HMM Clustering (MS-VBx) for Combined End-to-End and Vector Clustering-based Diarization. INTERSPEECH 2023: 3477-3481 - [i56]Katerina Zmolíková, Marc Delcroix, Tsubasa Ochiai, Keisuke Kinoshita, Jan Cernocký, Dong Yu:
Neural Target Speech Extraction: An Overview. CoRR abs/2301.13341 (2023) - [i55]Kohei Matsuura, Takanori Ashihara, Takafumi Moriya, Tomohiro Tanaka, Atsunori Ogawa, Marc Delcroix, Ryo Masumura:
Leveraging Large Text Corpora for End-to-End Speech Summarization. CoRR abs/2303.00978 (2023) - [i54]Marc Delcroix, Naohiro Tawara, Mireia Díez, Federico Landini, Anna Silnova, Atsunori Ogawa, Tomohiro Nakatani, Lukás Burget, Shoko Araki:
Multi-Stream Extension of Variational Bayesian HMM Clustering (MS-VBx) for Combined End-to-End and Vector Clustering-based Diarization. CoRR abs/2305.13580 (2023) - [i53]Hiroshi Sato, Ryo Masumura, Tsubasa Ochiai, Marc Delcroix, Takafumi Moriya, Takanori Ashihara, Kentaro Shinayama, Saki Mizuno, Mana Ihori, Tomohiro Tanaka, Nobukatsu Hojo:
Downstream Task Agnostic Speech Enhancement with Self-Supervised Representation Loss. CoRR abs/2305.14723 (2023) - [i52]Kohei Matsuura, Takanori Ashihara, Takafumi Moriya, Tomohiro Tanaka, Takatomo Kano, Atsunori Ogawa, Marc Delcroix:
Transfer Learning from Pre-trained Language Models Improves End-to-End Speech Summarization. CoRR abs/2306.04233 (2023) - [i51]Takanori Ashihara, Takafumi Moriya, Kohei Matsuura, Tomohiro Tanaka, Yusuke Ijima, Taichi Asami, Marc Delcroix, Yukinori Honma:
SpeechGLUE: How Well Can Self-Supervised Speech Models Capture Linguistic Knowledge? CoRR abs/2306.08374 (2023) - [i50]Thilo von Neumann, Christoph Böddeker, Marc Delcroix, Reinhold Haeb-Umbach:
MeetEval: A Toolkit for Computation of Word Error Rates for Meeting Transcription Systems. CoRR abs/2307.11394 (2023) - [i49]Naoyuki Kamo, Marc Delcroix, Tomohiro Nakatani:
Target Speech Extraction with Conditional Diffusion Model. CoRR abs/2308.03987 (2023) - [i48]Naohiro Tawara, Marc Delcroix, Atsushi Ando, Atsunori Ogawa:
NTT speaker diarization system for CHiME-7: multi-domain, multi-microphone End-to-end and vector clustering diarization. CoRR abs/2309.12656 (2023) - [i47]Thilo von Neumann, Christoph Böddeker, Tobias Cord-Landwehr, Marc Delcroix, Reinhold Haeb-Umbach:
Meeting Recognition with Continuous Speech Separation and Transcription-Supported Diarization. CoRR abs/2309.16482 (2023) - [i46]Dominik Klement, Mireia Díez, Federico Landini, Lukás Burget, Anna Silnova, Marc Delcroix, Naohiro Tawara:
Discriminative Training of VBx Diarization. CoRR abs/2310.02732 (2023) - [i45]Atsunori Ogawa, Takafumi Moriya, Naoyuki Kamo, Naohiro Tawara, Marc Delcroix:
Iterative Shallow Fusion of Backward Language Model for End-to-End Speech Recognition. CoRR abs/2310.11010 (2023) - [i44]Atsunori Ogawa, Naohiro Tawara, Marc Delcroix, Shoko Araki:
Lattice Rescoring Based on Large Ensemble of Complementary Neural Language Models. CoRR abs/2312.12764 (2023) - [i43]Atsunori Ogawa, Naohiro Tawara, Takatomo Kano, Marc Delcroix:
BLSTM-Based Confidence Estimation for End-to-End Speech Recognition. CoRR abs/2312.14609 (2023) - 2022
- [j24]Zili Huang, Marc Delcroix, Leibny Paola García-Perera, Shinji Watanabe, Desh Raj, Sanjeev Khudanpur:
Joint speaker diarization and speech recognition based on region proposal networks. Comput. Speech Lang. 72: 101316 (2022) - [c133]Thilo von Neumann, Keisuke Kinoshita, Christoph Böddeker, Marc Delcroix, Reinhold Haeb-Umbach:
SA-SDR: A Novel Loss Function for Separation of Meeting Style Data. ICASSP 2022: 6022-6026 - [c132]Takatomo Kano, Atsunori Ogawa, Marc Delcroix, Shinji Watanabe:
Integrating Multiple ASR Systems into NLP Backend with Attention Fusion. ICASSP 2022: 6237-6241 - [c131]Hiroshi Sato, Tsubasa Ochiai, Marc Delcroix, Keisuke Kinoshita, Naoyuki Kamo, Takafumi Moriya:
Learning to Enhance or Not: Neural Network-Based Switching of Enhanced and Observed Signals for Overlapping Speech Recognition. ICASSP 2022: 6287-6291 - [c130]Atsunori Ogawa, Naohiro Tawara, Marc Delcroix, Shoko Araki:
Lattice Rescoring Based on Large Ensemble of Complementary Neural Language Models. ICASSP 2022: 6517-6521 - [c129]Takafumi Moriya, Takanori Ashihara, Atsushi Ando, Hiroshi Sato, Tomohiro Tanaka, Kohei Matsuura, Ryo Masumura, Marc Delcroix, Takahiro Shinozaki:
Hybrid RNN-T/Attention-Based Streaming ASR with Triggered Chunkwise Attention and Dual Internal Language Model Integration. ICASSP 2022: 8282-8286 - [c128]Keisuke Kinoshita, Marc Delcroix, Tomoharu Iwata:
Tight Integration Of Neural- And Clustering-Based Diarization Through Deep Unfolding Of Infinite Gaussian Mixture Model. ICASSP 2022: 8382-8386 - [c127]Marc Delcroix, Keisuke Kinoshita, Tsubasa Ochiai, Katerina Zmolíková, Hiroshi Sato, Tomohiro Nakatani:
Listen only to me! How well can target speech extraction handle false alarms? INTERSPEECH 2022: 216-220 - [c126]Hiroshi Sato, Tsubasa Ochiai, Marc Delcroix, Keisuke Kinoshita, Takafumi Moriya, Naoki Makishima, Mana Ihori, Tomohiro Tanaka, Ryo Masumura:
Strategies to Improve Robustness of Target Speech Extraction to Enrollment Variations. INTERSPEECH 2022: 996-1000 - [c125]Keisuke Kinoshita, Thilo von Neumann, Marc Delcroix, Christoph Böddeker, Reinhold Haeb-Umbach:
Utterance-by-utterance overlap-aware neural diarization with Graph-PIT. INTERSPEECH 2022: 1486-1490 - [c124]Takafumi Moriya, Hiroshi Sato, Tsubasa Ochiai, Marc Delcroix, Takahiro Shinozaki:
Streaming Target-Speaker ASR with Neural Transducer. INTERSPEECH 2022: 2673-2677 - [c123]Martin Kocour, Katerina Zmolíková, Lucas Ondel, Jan Svec, Marc Delcroix, Tsubasa Ochiai, Lukás Burget, Jan Cernocký:
Revisiting joint decoding based multi-talker speech recognition with DNN acoustic model. INTERSPEECH 2022: 4955-4959 - [c122]Kazuma Iwamoto, Tsubasa Ochiai, Marc Delcroix, Rintaro Ikeshita, Hiroshi Sato, Shoko Araki, Shigeru Katagiri:
How bad are artifacts?: Analyzing the impact of speech enhancement errors on ASR. INTERSPEECH 2022: 5418-5422 - [c121]Jan Svec, Katerina Zmolíková, Martin Kocour, Marc Delcroix, Tsubasa Ochiai, Ladislav Mosner, Jan Honza Cernocký:
Analysis of Impact of Emotions on Target Speech Extraction and Speech Separation. IWAENC 2022: 1-5 - [c120]Yasunori Ohishi, Marc Delcroix, Tsubasa Ochiai, Shoko Araki, Daiki Takeuchi, Daisuke Niizumi, Akisato Kimura, Noboru Harada, Kunio Kashino:
ConceptBeam: Concept Driven Target Speech Extraction. ACM Multimedia 2022: 4252-4260 - [i42]Hiroshi Sato, Tsubasa Ochiai, Marc Delcroix, Keisuke Kinoshita, Naoyuki Kamo, Takafumi Moriya:
Learning to Enhance or Not: Neural Network-Based Switching of Enhanced and Observed Signals for Overlapping Speech Recognition. CoRR abs/2201.03881 (2022) - [i41]Kazuma Iwamoto, Tsubasa Ochiai, Marc Delcroix, Rintaro Ikeshita, Hiroshi Sato, Shoko Araki, Shigeru Katagiri:
How Bad Are Artifacts?: Analyzing the Impact of Speech Enhancement Errors on ASR. CoRR abs/2201.06685 (2022) - [i40]Keisuke Kinoshita, Marc Delcroix, Tomoharu Iwata:
Tight integration of neural- and clustering-based diarization through deep unfolding of infinite Gaussian mixture model. CoRR abs/2202.06524 (2022) - [i39]Marc Delcroix, Jorge Bennasar Vázquez, Tsubasa Ochiai, Keisuke Kinoshita, Yasunori Ohishi, Shoko Araki:
SoundBeam: Target sound extraction conditioned on sound-class labels and enrollment clues for increased performance and continuous learning. CoRR abs/2204.03895 (2022) - [i38]Marc Delcroix, Keisuke Kinoshita, Tsubasa Ochiai, Katerina Zmolíková, Hiroshi Sato, Tomohiro Nakatani:
Listen only to me! How well can target speech extraction handle false alarms? CoRR abs/2204.04811 (2022) - [i37]Tsubasa Ochiai, Marc Delcroix, Tomohiro Nakatani, Shoko Araki:
Mask-based Neural Beamforming for Moving Speakers with Self-Attention-based Tracking. CoRR abs/2205.03568 (2022) - [i36]Hiroshi Sato, Tsubasa Ochiai, Marc Delcroix, Keisuke Kinoshita, Takafumi Moriya, Naoki Makishima, Mana Ihori, Tomohiro Tanaka, Ryo Masumura:
Strategies to Improve Robustness of Target Speech Extraction to Enrollment Variations. CoRR abs/2206.08174 (2022) - [i35]Yasunori Ohishi, Marc Delcroix, Tsubasa Ochiai, Shoko Araki, Daiki Takeuchi, Daisuke Niizumi, Akisato Kimura, Noboru Harada, Kunio Kashino:
ConceptBeam: Concept Driven Target Speech Extraction. CoRR abs/2207.11964 (2022) - [i34]Keisuke Kinoshita, Thilo von Neumann, Marc Delcroix, Christoph Böddeker, Reinhold Haeb-Umbach:
Utterance-by-utterance overlap-aware neural diarization with Graph-PIT. CoRR abs/2207.13888 (2022) - [i33]Jan Svec, Katerina Zmolíková, Martin Kocour, Marc Delcroix, Tsubasa Ochiai, Ladislav Mosner, Jan Cernocký:
Analysis of impact of emotions on target speech extraction and speech separation. CoRR abs/2208.07091 (2022) - [i32]Takafumi Moriya, Hiroshi Sato, Tsubasa Ochiai, Marc Delcroix, Takahiro Shinozaki:
Streaming Target-Speaker ASR with Neural Transducer. CoRR abs/2209.04175 (2022) - [i31]Thilo von Neumann, Christoph Böddeker, Keisuke Kinoshita, Marc Delcroix, Reinhold Haeb-Umbach:
On Word Error Rate Definitions and their Efficient Computation for Multi-Speaker Speech Recognition Systems. CoRR abs/2211.16112 (2022) - 2021
- [j23]Reinhold Haeb-Umbach, Jahn Heymann, Lukas Drude, Shinji Watanabe, Marc Delcroix, Tomohiro Nakatani:
Far-Field Automatic Speech Recognition. Proc. IEEE 109(2): 124-148 (2021) - [c119]Thilo von Neumann, Christoph Böddeker, Keisuke Kinoshita, Marc Delcroix, Reinhold Haeb-Umbach:
Speeding Up Permutation Invariant Training for Source Separation. ITG Conference on Speech Communication 2021: 1-5 - [c118]Takatomo Kano, Atsunori Ogawa, Marc Delcroix, Shinji Watanabe:
Attention-Based Multi-Hypothesis Fusion for Speech Summarization. ASRU 2021: 487-494 - [c117]Julio Wissing, Benedikt T. Boenninghoff, Dorothea Kolossa, Tsubasa Ochiai, Marc Delcroix, Keisuke Kinoshita, Tomohiro Nakatani, Shoko Araki, Christopher Schymura:
Data Fusion for Audiovisual Speaker Localization: Extending Dynamic Stream Weights to the Spatial Domain. ICASSP 2021: 4705-4709 - [c116]Chenda Li, Zhuo Chen, Yi Luo, Cong Han, Tianyan Zhou, Keisuke Kinoshita, Marc Delcroix, Shinji Watanabe, Yanmin Qian:
Dual-Path Modeling for Long Recording Speech Separation in Meetings. ICASSP 2021: 5739-5743 - [c115]Marc Delcroix, Katerina Zmolíková, Tsubasa Ochiai, Keisuke Kinoshita, Tomohiro Nakatani:
Speaker Activity Driven Neural Speech Extraction. ICASSP 2021: 6099-6103 - [c114]Tsubasa Ochiai, Marc Delcroix, Tomohiro Nakatani, Rintaro Ikeshita, Keisuke Kinoshita, Shoko Araki:
Neural Network-Based Virtual Microphone Estimator. ICASSP 2021: 6114-6118 - [c113]Atsunori Ogawa, Naohiro Tawara, Takatomo Kano, Marc Delcroix:
BLSTM-Based Confidence Estimation for End-to-End Speech Recognition. ICASSP 2021: 6383-6387 - [c112]Wangyou Zhang, Christoph Böddeker, Shinji Watanabe, Tomohiro Nakatani, Marc Delcroix, Keisuke Kinoshita, Tsubasa Ochiai, Naoyuki Kamo, Reinhold Haeb-Umbach, Yanmin Qian:
End-to-End Dereverberation, Beamforming, and Speech Recognition with Improved Numerical Stability and Advanced Frontend. ICASSP 2021: 6898-6902 - [c111]Keisuke Kinoshita, Marc Delcroix, Naohiro Tawara:
Integrating End-to-End Neural and Clustering-Based Diarization: Getting the Best of Both Worlds. ICASSP 2021: 7198-7202 - [c110]Christoph Böddeker, Wangyou Zhang, Tomohiro Nakatani, Keisuke Kinoshita, Tsubasa Ochiai, Marc Delcroix, Naoyuki Kamo, Yanmin Qian, Reinhold Haeb-Umbach:
Convolutive Transfer Function Invariant SDR Training Criteria for Multi-Channel Reverberant Speech Separation. ICASSP 2021: 8428-8432 - [c109]Hiroshi Sato, Tsubasa Ochiai, Marc Delcroix, Keisuke Kinoshita, Takafumi Moriya, Naoyuki Kamo:
Should We Always Separate?: Switching Between Enhanced and Observed Signals for Overlapping Speech Recognition. Interspeech 2021: 1149-1153 - [c108]Katerina Zmolíková, Marc Delcroix, Desh Raj, Shinji Watanabe, Jan Cernocký:
Auxiliary Loss Function for Target Speech Extraction and Recognition with Weak Supervision Based on Speaker Characteristics. Interspeech 2021: 1464-1468 - [c107]Takafumi Moriya, Tomohiro Tanaka, Takanori Ashihara, Tsubasa Ochiai, Hiroshi Sato, Atsushi Ando, Ryo Masumura, Marc Delcroix, Taichi Asami:
Streaming End-to-End Speech Recognition for Hybrid RNN-T/Attention Architecture. Interspeech 2021: 1787-1791 - [c106]Christopher Schymura, Benedikt T. Bönninghoff, Tsubasa Ochiai, Marc Delcroix, Keisuke Kinoshita, Tomohiro Nakatani, Shoko Araki, Dorothea Kolossa:
PILOT: Introducing Transformers for Probabilistic Sound Event Localization. Interspeech 2021: 2117-2121 - [c105]Cong Han, Yi Luo, Chenda Li, Tianyan Zhou, Keisuke Kinoshita, Shinji Watanabe, Marc Delcroix, Hakan Erdogan, John R. Hershey, Nima Mesgarani, Zhuo Chen:
Continuous Speech Separation Using Speaker Inventory for Long Recording. Interspeech 2021: 3036-3040 - [c104]Thilo von Neumann, Keisuke Kinoshita, Christoph Böddeker, Marc Delcroix, Reinhold Haeb-Umbach:
Graph-PIT: Generalized Permutation Invariant Training for Continuous Separation of Arbitrary Numbers of Speakers. Interspeech 2021: 3490-3494 - [c103]Marc Delcroix, Jorge Bennasar Vázquez, Tsubasa Ochiai, Keisuke Kinoshita, Shoko Araki:
Few-Shot Learning of New Sound Classes for Target Sound Extraction. Interspeech 2021: 3500-3504 - [c102]Keisuke Kinoshita, Marc Delcroix, Naohiro Tawara:
Advances in Integration of End-to-End Neural and Clustering-Based Diarization for Real Conversational Speech. Interspeech 2021: 3565-3569 - [c101]Hiroshi Sato, Tsubasa Ochiai, Keisuke Kinoshita, Marc Delcroix, Tomohiro Nakatani, Shoko Araki:
Multimodal Attention Fusion for Target Speaker Extraction. SLT 2021: 778-784 - [c100]Chenda Li, Yi Luo, Cong Han, Jinyu Li, Takuya Yoshioka, Tianyan Zhou, Marc Delcroix, Keisuke Kinoshita, Christoph Böddeker, Yanmin Qian, Shinji Watanabe, Zhuo Chen:
Dual-Path RNN for Long Recording Speech Separation. SLT 2021: 865-872 - [c99]Katerina Zmolíková, Marc Delcroix, Lukás Burget, Tomohiro Nakatani, Jan Honza Cernocký:
Integration of Variational Autoencoder and Spatial Clustering for Adaptive Multi-Channel Neural Speech Separation. SLT 2021: 889-896 - [i30]Tsubasa Ochiai, Marc Delcroix, Tomohiro Nakatani, Rintaro Ikeshita, Keisuke Kinoshita, Shoko Araki:
Neural Network-based Virtual Microphone Estimator. CoRR abs/2101.04315 (2021) - [i29]Marc Delcroix, Katerina Zmolíková, Tsubasa Ochiai, Keisuke Kinoshita, Tomohiro Nakatani:
Speaker activity driven neural speech extraction. CoRR abs/2101.05516 (2021) - [i28]Hiroshi Sato, Tsubasa Ochiai, Keisuke Kinoshita, Marc Delcroix, Tomohiro Nakatani, Shoko Araki:
Multimodal Attention Fusion for Target Speaker Extraction. CoRR abs/2102.01326 (2021) - [i27]Wangyou Zhang, Christoph Böddeker, Shinji Watanabe, Tomohiro Nakatani, Marc Delcroix, Keisuke Kinoshita, Tsubasa Ochiai, Naoyuki Kamo, Reinhold Haeb-Umbach, Yanmin Qian:
End-to-End Dereverberation, Beamforming, and Speech Recognition with Improved Numerical Stability and Advanced Frontend. CoRR abs/2102.11525 (2021) - [i26]Julio Wissing, Benedikt T. Boenninghoff, Dorothea Kolossa, Tsubasa Ochiai, Marc Delcroix, Keisuke Kinoshita, Tomohiro Nakatani, Shoko Araki, Christopher Schymura:
Data Fusion for Audiovisual Speaker Localization: Extending Dynamic Stream Weights to the Spatial Domain. CoRR abs/2102.11588 (2021) - [i25]Chenda Li, Zhuo Chen, Yi Luo, Cong Han, Tianyan Zhou, Keisuke Kinoshita, Marc Delcroix, Shinji Watanabe, Yanmin Qian:
Dual-Path Modeling for Long Recording Speech Separation in Meetings. CoRR abs/2102.11634 (2021) - [i24]Christopher Schymura, Tsubasa Ochiai, Marc Delcroix, Keisuke Kinoshita, Tomohiro Nakatani, Shoko Araki, Dorothea Kolossa:
Exploiting Attention-based Sequence-to-Sequence Architectures for Sound Event Localization. CoRR abs/2103.00417 (2021) - [i23]Keisuke Kinoshita, Marc Delcroix, Naohiro Tawara:
Advances in integration of end-to-end neural and clustering-based diarization for real conversational speech. CoRR abs/2105.09040 (2021) - [i22]Hiroshi Sato, Tsubasa Ochiai, Marc Delcroix, Keisuke Kinoshita, Takafumi Moriya, Naoyuki Kamo:
Should We Always Separate?: Switching Between Enhanced and Observed Signals for Overlapping Speech Recognition. CoRR abs/2106.00949 (2021) - [i21]Christopher Schymura, Benedikt T. Bönninghoff, Tsubasa Ochiai, Marc Delcroix, Keisuke Kinoshita, Tomohiro Nakatani, Shoko Araki, Dorothea Kolossa:
PILOT: Introducing Transformers for Probabilistic Sound Event Localization. CoRR abs/2106.03903 (2021) - [i20]Marc Delcroix, Jorge Bennasar Vázquez, Tsubasa Ochiai, Keisuke Kinoshita, Shoko Araki:
Few-shot learning of new sound classes for target sound extraction. CoRR abs/2106.07144 (2021) - [i19]Thilo von Neumann, Christoph Böddeker, Keisuke Kinoshita, Marc Delcroix, Reinhold Haeb-Umbach:
Speeding Up Permutation Invariant Training for Source Separation. CoRR abs/2107.14445 (2021) - [i18]Thilo von Neumann, Keisuke Kinoshita, Christoph Böddeker, Marc Delcroix, Reinhold Haeb-Umbach:
Graph-PIT: Generalized permutation invariant training for continuous separation of arbitrary numbers of speakers. CoRR abs/2107.14446 (2021) - [i17]Thilo von Neumann, Keisuke Kinoshita, Christoph Böddeker, Marc Delcroix, Reinhold Haeb-Umbach:
SA-SDR: A novel loss function for separation of meeting style data. CoRR abs/2110.15581 (2021) - [i16]Martin Kocour, Katerina Zmolíková, Lucas Ondel, Jan Svec, Marc Delcroix, Tsubasa Ochiai, Lukás Burget, Jan Cernocký:
Revisiting joint decoding based multi-talker speech recognition with DNN acoustic model. CoRR abs/2111.00009 (2021) - [i15]Takatomo Kano, Atsunori Ogawa, Marc Delcroix, Shinji Watanabe:
Attention-based Multi-hypothesis Fusion for Speech Summarization. CoRR abs/2111.08201 (2021) - 2020
- [j22]Tomohiro Nakatani, Christoph Böddeker, Keisuke Kinoshita, Rintaro Ikeshita, Marc Delcroix, Reinhold Haeb-Umbach:
Jointly Optimal Denoising, Dereverberation, and Source Separation. IEEE ACM Trans. Audio Speech Lang. Process. 28: 2267-2282 (2020) - [c98]Christopher Schymura, Tsubasa Ochiai, Marc Delcroix, Keisuke Kinoshita, Tomohiro Nakatani, Shoko Araki, Dorothea Kolossa:
Exploiting Attention-based Sequence-to-Sequence Architectures for Sound Event Localization. EUSIPCO 2020: 231-235 - [c97]Yuma Koizumi, Kohei Yatabe, Marc Delcroix, Yoshiki Masuyama, Daiki Takeuchi:
Speech Enhancement Using Self-Adaptation and Multi-Head Self-Attention. ICASSP 2020: 181-185 - [c96]Keisuke Kinoshita, Marc Delcroix, Shoko Araki, Tomohiro Nakatani:
Tackling Real Noisy Reverberant Meetings with All-Neural Source Separation, Counting, and Diarization System. ICASSP 2020: 381-385 - [c95]Christopher Schymura, Tsubasa Ochiai, Marc Delcroix, Keisuke Kinoshita, Tomohiro Nakatani, Shoko Araki, Dorothea Kolossa:
A Dynamic Stream Weight Backprop Kalman Filter for Audiovisual Speaker Tracking. ICASSP 2020: 581-585 - [c94]Marc Delcroix, Tsubasa Ochiai, Katerina Zmolíková, Keisuke Kinoshita, Naohiro Tawara, Tomohiro Nakatani, Shoko Araki:
Improving Speaker Discrimination of Target Speech Extraction With Time-Domain Speakerbeam. ICASSP 2020: 691-695 - [c93]Tsubasa Ochiai, Marc Delcroix, Rintaro Ikeshita, Keisuke Kinoshita, Tomohiro Nakatani, Shoko Araki:
Beam-TasNet: Time-domain Audio Separation Network Meets Frequency-domain Beamformer. ICASSP 2020: 6384-6388 - [c92]Tomohiro Nakatani, Riki Takahashi, Tsubasa Ochiai, Keisuke Kinoshita, Rintaro Ikeshita, Marc Delcroix, Shoko Araki:
DNN-supported Mask-based Convolutional Beamforming for Simultaneous Denoising, Dereverberation, and Source Separation. ICASSP 2020: 6399-6403 - [c91]Naohiro Tawara, Atsunori Ogawa, Tomoharu Iwata, Marc Delcroix, Tetsuji Ogawa:
Frame-Level Phoneme-Invariant Speaker Embedding for Text-Independent Speaker Recognition on Extremely Short Utterances. ICASSP 2020: 6799-6803 - [c90]Thilo von Neumann, Keisuke Kinoshita, Lukas Drude, Christoph Böddeker, Marc Delcroix, Tomohiro Nakatani, Reinhold Haeb-Umbach:
End-to-End Training of Time Domain Audio Separation and Recognition. ICASSP 2020: 7004-7008 - [c89]Keisuke Kinoshita, Tsubasa Ochiai, Marc Delcroix, Tomohiro Nakatani:
Improving Noise Robust Automatic Speech Recognition with Single-Channel Time-Domain Enhancement Network. ICASSP 2020: 7009-7013 - [c88]Takafumi Moriya, Tsubasa Ochiai, Shigeki Karita, Hiroshi Sato, Tomohiro Tanaka, Takanori Ashihara, Ryo Masumura, Yusuke Shinohara, Marc Delcroix:
Self-Distillation for Improving CTC-Transformer-Based ASR Systems. INTERSPEECH 2020: 546-550 - [c87]Tsubasa Ochiai, Marc Delcroix, Yuma Koizumi, Hiroaki Ito, Keisuke Kinoshita, Shoko Araki:
Listen to What You Want: Neural Network-Based Universal Sound Selector. INTERSPEECH 2020: 1441-1445 - [c86]Keisuke Kinoshita, Thilo von Neumann, Marc Delcroix, Tomohiro Nakatani, Reinhold Haeb-Umbach:
Multi-Path RNN for Hierarchical Modeling of Long Sequential Data and its Application to Speaker Stream Separation. INTERSPEECH 2020: 2652-2656 - [c85]Thilo von Neumann, Christoph Böddeker, Lukas Drude, Keisuke Kinoshita, Marc Delcroix, Tomohiro Nakatani, Reinhold Haeb-Umbach:
Multi-Talker ASR for an Unknown Number of Sources: Joint Training of Source Counting, Separation and ASR. INTERSPEECH 2020: 3097-3101 - [c84]Atsunori Ogawa, Naohiro Tawara, Marc Delcroix:
Language Model Data Augmentation Based on Text Domain Transfer. INTERSPEECH 2020: 4926-4930 - [c83]Ali Aroudi, Marc Delcroix, Tomohiro Nakatani, Keisuke Kinoshita, Shoko Araki, Simon Doclo:
Cognitive-Driven Convolutional Beamforming Using EEG-Based Auditory Attention Decoding. MLSP 2020: 1-6 - [i14]Marc Delcroix, Tsubasa Ochiai, Katerina Zmolíková, Keisuke Kinoshita, Naohiro Tawara, Tomohiro Nakatani, Shoko Araki:
Improving speaker discrimination of target speech extraction with time-domain SpeakerBeam. CoRR abs/2001.08378 (2020) - [i13]Yuma Koizumi, Kohei Yatabe, Marc Delcroix, Yoshiki Masuyama, Daiki Takeuchi:
Speech Enhancement using Self-Adaptation and Multi-Head Self-Attention. CoRR abs/2002.05873 (2020) - [i12]Keisuke Kinoshita, Marc Delcroix, Shoko Araki, Tomohiro Nakatani:
Tackling real noisy reverberant meetings with all-neural source separation, counting, and diarization system. CoRR abs/2003.03987 (2020) - [i11]Keisuke Kinoshita, Tsubasa Ochiai, Marc Delcroix, Tomohiro Nakatani:
Improving noise robust automatic speech recognition with single-channel time-domain enhancement network. CoRR abs/2003.03998 (2020) - [i10]Ali Aroudi, Marc Delcroix, Tomohiro Nakatani, Keisuke Kinoshita, Shoko Araki, Simon Doclo:
Cognitive-driven convolutional beamforming using EEG-based auditory attention decoding. CoRR abs/2005.04669 (2020) - [i9]Tomohiro Nakatani, Christoph Böddeker, Keisuke Kinoshita, Rintaro Ikeshita, Marc Delcroix, Reinhold Haeb-Umbach:
Jointly optimal denoising, dereverberation, and source separation. CoRR abs/2005.09843 (2020) - [i8]Thilo von Neumann, Christoph Böddeker, Lukas Drude, Keisuke Kinoshita, Marc Delcroix, Tomohiro Nakatani, Reinhold Haeb-Umbach:
Multi-talker ASR for an unknown number of sources: Joint training of source counting, separation and ASR. CoRR abs/2006.02786 (2020) - [i7]Tsubasa Ochiai, Marc Delcroix, Yuma Koizumi, Hiroaki Ito, Keisuke Kinoshita, Shoko Araki:
Listen to What You Want: Neural Network-based Universal Sound Selector. CoRR abs/2006.05712 (2020) - [i6]Keisuke Kinoshita, Thilo von Neumann, Marc Delcroix, Tomohiro Nakatani, Reinhold Haeb-Umbach:
Multi-path RNN for hierarchical modeling of long sequential data and its application to speaker stream separation. CoRR abs/2006.13579 (2020) - [i5]Keisuke Kinoshita, Marc Delcroix, Naohiro Tawara:
Integrating end-to-end neural and clustering-based diarization: Getting the best of both worlds. CoRR abs/2010.13366 (2020) - [i4]Christoph Böddeker, Wangyou Zhang, Tomohiro Nakatani, Keisuke Kinoshita, Tsubasa Ochiai, Marc Delcroix, Naoyuki Kamo, Yanmin Qian, Shinji Watanabe, Reinhold Haeb-Umbach:
Convolutive Transfer Function Invariant SDR training criteria for Multi-Channel Reverberant Speech Separation. CoRR abs/2011.15003 (2020) - [i3]Cong Han, Yi Luo, Chenda Li, Tianyan Zhou, Keisuke Kinoshita, Shinji Watanabe, Marc Delcroix, Hakan Erdogan, John R. Hershey, Nima Mesgarani, Zhuo Chen:
Continuous Speech Separation Using Speaker Inventory for Long Multi-talker Recording. CoRR abs/2012.09727 (2020)
2010 – 2019
- 2019
- [j21]Michael Hentschel, Marc Delcroix, Atsunori Ogawa, Tomoharu Iwata, Tomohiro Nakatani:
Feature Based Domain Adaptation for Neural Network Language Models with Factorised Hidden Layers. IEICE Trans. Inf. Syst. 102-D(3): 598-608 (2019) - [j20]Katerina Zmolíková, Marc Delcroix, Keisuke Kinoshita, Tsubasa Ochiai, Tomohiro Nakatani, Lukás Burget, Jan Cernocký:
SpeakerBeam: Speaker Aware Neural Network for Target Speaker Extraction in Speech Mixtures. IEEE J. Sel. Top. Signal Process. 13(4): 800-814 (2019) - [c82]Shoko Araki, Nobutaka Ono, Keisuke Kinoshita, Marc Delcroix:
Projection Back onto Filtered Observations for Speech Separation with Distributed Microphone Array. CAMSAP 2019: 291-295 - [c81]Thilo von Neumann, Keisuke Kinoshita, Marc Delcroix, Shoko Araki, Tomohiro Nakatani, Reinhold Haeb-Umbach:
All-neural Online Source Separation, Counting, and Diarization for Meeting Analysis. ICASSP 2019: 91-95 - [c80]Shoko Araki, Nobutaka Ono, Keisuke Kinoshita, Marc Delcroix:
Estimation of Sampling Frequency Mismatch between Distributed Asynchronous Microphones under Existence of Source Movements with Stationary Time Periods Detection. ICASSP 2019: 785-789 - [c79]Shigeki Karita, Shinji Watanabe, Tomoharu Iwata, Marc Delcroix, Atsunori Ogawa, Tomohiro Nakatani:
Semi-supervised End-to-end Speech Recognition Using Text-to-speech and Autoencoders. ICASSP 2019: 6166-6170 - [c78]Yuki Kubo, Tomohiro Nakatani, Marc Delcroix, Keisuke Kinoshita, Shoko Araki:
Mask-based MVDR Beamformer for Noisy Multisource Environments: Introduction of Time-varying Spatial Covariance Model. ICASSP 2019: 6855-6859 - [c77]Marc Delcroix, Katerina Zmolíková, Tsubasa Ochiai, Keisuke Kinoshita, Shoko Araki, Tomohiro Nakatani:
Compact Network for Speakerbeam Target Speaker Extraction. ICASSP 2019: 6965-6969 - [c76]Tsubasa Ochiai, Marc Delcroix, Keisuke Kinoshita, Atsunori Ogawa, Tomohiro Nakatani:
A Unified Framework for Neural Speech Separation and Extraction. ICASSP 2019: 6975-6979 - [c75]Michael Hentschel, Marc Delcroix, Atsunori Ogawa, Tomoharu Iwata, Tomohiro Nakatani:
A Unified Framework for Feature-based Domain Adaptation of Neural Network Language Models. ICASSP 2019: 7250-7254 - [c74]Marc Delcroix, Shinji Watanabe, Tsubasa Ochiai, Keisuke Kinoshita, Shigeki Karita, Atsunori Ogawa, Tomohiro Nakatani:
End-to-End SpeakerBeam for Single Channel Target Speech Recognition. INTERSPEECH 2019: 451-455 - [c73]Shigeki Karita, Nelson Enrique Yalta Soplin, Shinji Watanabe, Marc Delcroix, Atsunori Ogawa, Tomohiro Nakatani:
Improving Transformer-Based End-to-End Speech Recognition with Connectionist Temporal Classification and Language Model Integration. INTERSPEECH 2019: 1408-1412 - [c72]Tsubasa Ochiai, Marc Delcroix, Keisuke Kinoshita, Atsunori Ogawa, Tomohiro Nakatani:
Multimodal SpeakerBeam: Single Channel Target Speech Extraction with Audio-Visual Speaker Clues. INTERSPEECH 2019: 2718-2722 - [c71]Atsunori Ogawa, Marc Delcroix, Shigeki Karita, Tomohiro Nakatani:
Improved Deep Duel Model for Rescoring N-Best Speech Recognition List Using Backward LSTMLM and Ensemble Encoders. INTERSPEECH 2019: 3900-3904 - [i2]Thilo von Neumann, Keisuke Kinoshita, Marc Delcroix, Shoko Araki, Tomohiro Nakatani, Reinhold Haeb-Umbach:
All-neural online source separation, counting, and diarization for meeting analysis. CoRR abs/1902.07881 (2019) - [i1]Thilo von Neumann, Keisuke Kinoshita, Lukas Drude, Christoph Böddeker, Marc Delcroix, Tomohiro Nakatani, Reinhold Haeb-Umbach:
End-to-end training of time domain audio separation and recognition. CoRR abs/1912.08462 (2019) - 2018
- [j19]Marc Delcroix, Keisuke Kinoshita, Atsunori Ogawa, Christian Huemmer, Tomohiro Nakatani:
Context Adaptive Neural Network Based Acoustic Models for Rapid Adaptation. IEEE ACM Trans. Audio Speech Lang. Process. 26(5): 895-908 (2018) - [c70]Takafumi Moriya, Ryo Masumura, Taichi Asami, Yusuke Shinohara, Marc Delcroix, Yoshikazu Yamaguchi, Yushi Aono:
Progressive Neural Network-based Knowledge Transfer in Acoustic Models. APSIPA 2018: 998-1002 - [c69]Michael Hentschel, Marc Delcroix, Atsunori Ogawa, Tomohiro Nakatani:
Feature-Based Learning Hidden Unit Contributions for Domain Adaptation of RNN-LMs. APSIPA 2018: 1692-1696 - [c68]Michael Hentschel, Marc Delcroix, Atsunori Ogawa, Tomoharu Iwata, Tomohiro Nakatani:
Factorised Hidden Layer Based Domain Adaptation for Recurrent Neural Network Language Models. APSIPA 2018: 1940-1944 - [c67]Keisuke Kinoshita, Lukas Drude, Marc Delcroix, Tomohiro Nakatani:
Listening to Each Speaker One by One with Recurrent Selective Hearing Networks. ICASSP 2018: 5064-5068 - [c66]Marc Delcroix, Katerina Zmolíková, Keisuke Kinoshita, Atsunori Ogawa, Tomohiro Nakatani:
Single Channel Target Speaker Extraction and Recognition with Speaker Beam. ICASSP 2018: 5554-5558 - [c65]Shoko Araki, Nobutaka Ono, Keisuke Kinoshita, Marc Delcroix:
Meeting Recognition with Asynchronous Distributed Microphone Array Using Block-Wise Refinement of Mask-Based MVDR Beamformer. ICASSP 2018: 5694-5698 - [c64]Shigeki Karita, Atsunori Ogawa, Marc Delcroix, Tomohiro Nakatani:
Sequence Training of Encoder-Decoder Model Using Policy Gradient for End-to-End Speech Recognition. ICASSP 2018: 5839-5843 - [c63]Atsunori Ogawa, Marc Delcroix, Shigeki Karita, Tomohiro Nakatani:
Rescoring N-Best Speech Recognition List Based on One-on-One Hypothesis Comparison Using Encoder-Classifier Model. ICASSP 2018: 6099-6103 - [c62]Katerina Zmolíková, Marc Delcroix, Keisuke Kinoshita, Takuya Higuchi, Tomohiro Nakatani, Jan Cernocký:
Optimization of Speaker-Aware Multichannel Speech Extraction with ASR Criterion. ICASSP 2018: 6702-6706 - [c61]Shigeki Karita, Shinji Watanabe, Tomoharu Iwata, Atsunori Ogawa, Marc Delcroix:
Semi-Supervised End-to-End Speech Recognition. INTERSPEECH 2018: 2-6 - [c60]Takafumi Moriya, Sei Ueno, Yusuke Shinohara, Marc Delcroix, Yoshikazu Yamaguchi, Yushi Aono:
Multi-task Learning with Augmentation Strategy for Acoustic-to-word Attention-based Encoder-decoder Speech Recognition. INTERSPEECH 2018: 2399-2403 - [c59]Marc Delcroix, Shinji Watanabe, Atsunori Ogawa, Shigeki Karita, Tomohiro Nakatani:
Auxiliary Feature Based Adaptation of End-to-end ASR Systems. INTERSPEECH 2018: 2444-2448 - [c58]Lukas Drude, Christoph Böddeker, Jahn Heymann, Reinhold Haeb-Umbach, Keisuke Kinoshita, Marc Delcroix, Tomohiro Nakatani:
Integrating Neural Network Based Beamforming and Weighted Prediction Error Dereverberation. INTERSPEECH 2018: 3043-3047 - [c57]Yutaro Matsui, Tomohiro Nakatani, Marc Delcroix, Keisuke Kinoshita, Nobutaka Ito, Shoko Araki, Shoji Makino:
Online Integration of DNN-Based and Spatial Clustering-Based Mask Estimation for Robust MVDR Beamforming. IWAENC 2018: 71-75 - [c56]Shoko Araki, Nobutaka Ono, Keisuke Kinoshita, Marc Delcroix:
Comparison of Reference Microphone Selection Algorithms for Distributed Microphone Array Based Speech Enhancement in Meeting Recognition Scenarios. IWAENC 2018: 316-320 - 2017
- [j18]Takuya Higuchi, Nobutaka Ito, Shoko Araki, Takuya Yoshioka, Marc Delcroix, Tomohiro Nakatani:
Online MVDR Beamformer Based on Complex Gaussian Mixture Model With Spatial Prior for Noise Robust ASR. IEEE ACM Trans. Audio Speech Lang. Process. 25(4): 780-793 (2017) - [c55]Michael Hentschel, Atsunori Ogawa, Marc Delcroix, Tomohiro Nakatani, Yuji Matsumoto:
Exploiting imbalanced textual and acoustic data for training prosodically-enhanced RNNLMs. APSIPA 2017: 618-621 - [c54]Katerina Zmolíková, Marc Delcroix, Keisuke Kinoshita, Takuya Higuchi, Atsunori Ogawa, Tomohiro Nakatani:
Learning speaker representation for neural network based multichannel speaker extraction. ASRU 2017: 8-15 - [c53]Shoko Araki, Nobutaka Ono, Keisuke Kinoshita, Marc Delcroix:
Meeting recognition with asynchronous distributed microphone array. ASRU 2017: 32-39 - [c52]Takuya Higuchi, Keisuke Kinoshita, Marc Delcroix, Tomohiro Nakatani:
Adversarial training for data-driven speech enhancement without parallel corpus. ASRU 2017: 40-47 - [c51]Shoko Araki, Nobutaka Ito, Marc Delcroix, Atsunori Ogawa, Keisuke Kinoshita, Takuya Higuchi, Takuya Yoshioka, Dung T. Tran, Shigeki Karita, Tomohiro Nakatani:
Online meeting recognition in noisy environments with time-frequency mask based MVDR beamforming. HSCMA 2017: 16-20 - [c50]Keisuke Kinoshita, Marc Delcroix, Atsunori Ogawa, Takuya Higuchi, Tomohiro Nakatani:
Deep mixture density network for statistical model-based feature enhancement. ICASSP 2017: 251-255 - [c49]Nobutaka Ito, Shoko Araki, Marc Delcroix, Tomohiro Nakatani:
Probabilistic spatial dictionary based online adaptive beamforming for meeting recognition in noisy and reverberant environments. ICASSP 2017: 681-685 - [c48]Christian Huemmer, Marc Delcroix, Atsunori Ogawa, Keisuke Kinoshita, Tomohiro Nakatani, Walter Kellermann:
Online environmental adaptation of CNN-based acoustic models using spatial diffuseness features. ICASSP 2017: 4875-4879 - [c47]Tsubasa Ochiai, Marc Delcroix, Keisuke Kinoshita, Atsunori Ogawa, Taichi Asami, Shigeru Katagiri, Tomohiro Nakatani:
Cumulative moving averaged bottleneck speaker vectors for online speaker adaptation of CNN-based acoustic models. ICASSP 2017: 5175-5179 - [c46]Dung T. Tran, Marc Delcroix, Atsunori Ogawa, Christian Huemmer, Tomohiro Nakatani:
Feedback connection for deep neural network-based acoustic modeling. ICASSP 2017: 5240-5244 - [c45]Keisuke Kinoshita, Marc Delcroix, Haeyong Kwon, Takuma Mori, Tomohiro Nakatani:
Neural Network-Based Spectrum Estimation for Online WPE Dereverberation. INTERSPEECH 2017: 384-388 - [c44]Takuya Higuchi, Keisuke Kinoshita, Marc Delcroix, Katerina Zmolíková, Tomohiro Nakatani:
Deep Clustering-Based Beamforming for Separation with Unknown Number of Sources. INTERSPEECH 2017: 1183-1187 - [c43]Dung T. Tran, Marc Delcroix, Shigeki Karita, Michael Hentschel, Atsunori Ogawa, Tomohiro Nakatani:
Unfolded Deep Recurrent Convolutional Neural Network with Jump Ahead Connections for Acoustic Modeling. INTERSPEECH 2017: 1596-1600 - [c42]Shigeki Karita, Atsunori Ogawa, Marc Delcroix, Tomohiro Nakatani:
Forward-Backward Convolutional LSTM for Acoustic Modeling. INTERSPEECH 2017: 1601-1605 - [c41]Atsunori Ogawa, Keisuke Kinoshita, Marc Delcroix, Tomohiro Nakatani:
Improved Example-Based Speech Enhancement by Using Deep Neural Network Acoustic Model for Noise Robust Example Search. INTERSPEECH 2017: 1963-1967 - [c40]Katerina Zmolíková, Marc Delcroix, Keisuke Kinoshita, Takuya Higuchi, Atsunori Ogawa, Tomohiro Nakatani:
Speaker-Aware Neural Network Based Beamformer for Speaker Extraction in Speech Mixtures. INTERSPEECH 2017: 2655-2659 - [c39]Dung T. Tran, Marc Delcroix, Atsunori Ogawa, Tomohiro Nakatani:
Uncertainty Decoding with Adaptive Sampling for Noise Robust DNN-Based Acoustic Modeling. INTERSPEECH 2017: 3852-3856 - [p7]Shinji Watanabe, Marc Delcroix, Florian Metze, John R. Hershey:
Preliminaries. New Era for Robust Speech Recognition, Exploiting Deep Learning 2017: 3-17 - [p6]Marc Delcroix, Takuya Yoshioka, Nobutaka Ito, Atsunori Ogawa, Keisuke Kinoshita, Masakiyo Fujimoto, Takuya Higuchi, Shoko Araki, Tomohiro Nakatani:
Multichannel Speech Enhancement Approaches to DNN-Based Far-Field Speech Recognition. New Era for Robust Speech Recognition, Exploiting Deep Learning 2017: 21-49 - [p5]Martin Karafiát, Karel Veselý, Katerina Zmolíková, Marc Delcroix, Shinji Watanabe, Lukás Burget, Jan Honza Cernocký, Igor Szöke:
Training Data Augmentation and Data Selection. New Era for Robust Speech Recognition, Exploiting Deep Learning 2017: 245-260 - [p4]Keisuke Kinoshita, Marc Delcroix, Sharon Gannot, Emanuël A. P. Habets, Reinhold Haeb-Umbach, Walter Kellermann, Volker Leutnant, Roland Maas, Tomohiro Nakatani, Bhiksha Raj, Armin Sehr, Takuya Yoshioka:
The REVERB Challenge: A Benchmark Task for Reverberation-Robust ASR Techniques. New Era for Robust Speech Recognition, Exploiting Deep Learning 2017: 345-354 - [p3]Shinji Watanabe, Takaaki Hori, Yajie Miao, Marc Delcroix, Florian Metze, John R. Hershey:
Toolkits for Robust Speech Processing. New Era for Robust Speech Recognition, Exploiting Deep Learning 2017: 369-382 - [e1]Shinji Watanabe, Marc Delcroix, Florian Metze, John R. Hershey:
New Era for Robust Speech Recognition, Exploiting Deep Learning. Springer 2017, ISBN 978-3-319-64679-4 [contents] - 2016
- [j17]Marc Delcroix, Atsunori Ogawa, Seong-Jun Hahm, Tomohiro Nakatani, Atsushi Nakamura:
Differenced maximum mutual information criterion for robust unsupervised acoustic model adaptation. Comput. Speech Lang. 36: 24-41 (2016) - [j16]Keisuke Kinoshita, Marc Delcroix, Sharon Gannot, Emanuël A. P. Habets, Reinhold Haeb-Umbach, Walter Kellermann, Volker Leutnant, Roland Maas, Tomohiro Nakatani, Bhiksha Raj, Armin Sehr, Takuya Yoshioka:
A summary of the REVERB challenge: state-of-the-art and remaining challenges in reverberant speech processing research. EURASIP J. Adv. Signal Process. 2016: 7 (2016) - [c38]Souvik Kundu, Gautam Mantena, Yanmin Qian, Tian Tan, Marc Delcroix, Khe Chai Sim:
Joint acoustic factor learning for robust deep neural network based automatic speech recognition. ICASSP 2016: 5025-5029 - [c37]Marc Delcroix, Keisuke Kinoshita, Chengzhu Yu, Atsunori Ogawa, Takuya Yoshioka, Tomohiro Nakatani:
Context adaptive deep neural networks for fast acoustic model adaptation in noisy conditions. ICASSP 2016: 5270-5274 - [c36]Marc Delcroix, Keisuke Kinoshita, Atsunori Ogawa, Takuya Yoshioka, Dung T. Tran, Tomohiro Nakatani:
Context Adaptive Neural Network for Rapid Adaptation of Deep CNN Based Acoustic Models. INTERSPEECH 2016: 1573-1577 - [c35]Katerina Zmolíková, Martin Karafiát, Karel Veselý, Marc Delcroix, Shinji Watanabe, Lukás Burget, Jan Cernocký:
Data Selection by Sequence Summarizing Neural Network in Mismatch Condition Training. INTERSPEECH 2016: 2354-2358 - [c34]Atsunori Ogawa, Shogo Seki, Keisuke Kinoshita, Marc Delcroix, Takuya Yoshioka, Tomohiro Nakatani, Kazuya Takeda:
Robust Example Search Using Bottleneck Features for Example-Based Speech Enhancement. INTERSPEECH 2016: 3733-3737 - [c33]Dung T. Tran, Marc Delcroix, Atsunori Ogawa, Tomohiro Nakatani:
Factorized Linear Input Network for Acoustic Model Adaptation in Noisy Conditions. INTERSPEECH 2016: 3813-3817 - 2015
- [j15]Marc Delcroix, Takuya Yoshioka, Atsunori Ogawa, Yotaro Kubo, Masakiyo Fujimoto, Nobutaka Ito, Keisuke Kinoshita, Miquel Espi, Shoko Araki, Takaaki Hori, Tomohiro Nakatani:
Strategies for distant speech recognitionin reverberant environments. EURASIP J. Adv. Signal Process. 2015: 60 (2015) - [c32]Takuya Yoshioka, Nobutaka Ito, Marc Delcroix, Atsunori Ogawa, Keisuke Kinoshita, Masakiyo Fujimoto, Chengzhu Yu, Wojciech J. Fabian, Miquel Espi, Takuya Higuchi, Shoko Araki, Tomohiro Nakatani:
The NTT CHiME-3 system: Advances in speech enhancement and recognition for mobile multi-microphone devices. ASRU 2015: 436-443 - [c31]Shoko Araki, Tomoki Hayashi, Marc Delcroix, Masakiyo Fujimoto, Kazuya Takeda, Tomohiro Nakatani:
Exploring multi-channel features for denoising-autoencoder-based speech enhancement. ICASSP 2015: 116-120 - [c30]Marc Delcroix, Keisuke Kinoshita, Takaaki Hori, Tomohiro Nakatani:
Context adaptive deep neural networks for fast acoustic model adaptation. ICASSP 2015: 4535-4539 - [c29]Quoc Truong Do, Satoshi Nakamura, Marc Delcroix, Takaaki Hori:
WFST-based structural classification integrating dnn acoustic features and RNN language features for speech recognition. ICASSP 2015: 4959-4963 - [c28]Keisuke Kinoshita, Marc Delcroix, Atsunori Ogawa, Tomohiro Nakatani:
Text-informed speech enhancement with deep neural networks. INTERSPEECH 2015: 1760-1764 - [c27]Chengzhu Yu, Atsunori Ogawa, Marc Delcroix, Takuya Yoshioka, Tomohiro Nakatani, John H. L. Hansen:
Robust i-vector extraction for neural network adaptation in noisy environment. INTERSPEECH 2015: 2854-2857 - 2014
- [j14]Mehrez Souden, Keisuke Kinoshita, Marc Delcroix, Tomohiro Nakatani:
Location Feature Integration for Clustering-Based Speech Separation in Distributed Microphone Arrays. IEEE ACM Trans. Audio Speech Lang. Process. 22(2): 354-367 (2014) - [c26]Marc Delcroix, Takuya Yoshioka, Atsunori Ogawa, Yotaro Kubo, Masakiyo Fujimoto, Nobutaka Ito, Keisuke Kinoshita, Miquel Espi, Shoko Araki, Takaaki Hori, Tomohiro Nakatani:
Defeating reverberation: Advanced dereverberation and recognition techniques for hands-free speech recognition. GlobalSIP 2014: 522-526 - 2013
- [j13]Marc Delcroix, Shinji Watanabe, Tomohiro Nakatani, Atsushi Nakamura:
Cluster-based dynamic variance adaptation for interconnecting speech enhancement pre-processor and speech recognizer. Comput. Speech Lang. 27(1): 350-368 (2013) - [j12]Marc Delcroix, Keisuke Kinoshita, Tomohiro Nakatani, Shoko Araki, Atsunori Ogawa, Takaaki Hori, Shinji Watanabe, Masakiyo Fujimoto, Takuya Yoshioka, Takanobu Oba, Yotaro Kubo, Mehrez Souden, Seong-Jun Hahm, Atsushi Nakamura:
Speech recognition in living rooms: Integrated speech enhancement and recognition system based on spatial, spectral and temporal modeling of sounds. Comput. Speech Lang. 27(3): 851-873 (2013) - [j11]Tomohiro Nakatani, Shoko Araki, Takuya Yoshioka, Marc Delcroix, Masakiyo Fujimoto:
Dominance Based Integration of Spatial and Spectral Features for Speech Enhancement. IEEE ACM Trans. Audio Speech Lang. Process. 21(12): 2516-2531 (2013) - [c25]Marc Delcroix, Atsunori Ogawa, Seong-Jun Hahm, Tomohiro Nakatani, Atsushi Nakamura:
Unsupervised discriminative adaptation using differenced maximum mutual information based linear regression. ICASSP 2013: 7888-7892 - [c24]Seong-Jun Hahm, Atsunori Ogawa, Marc Delcroix, Masakiyo Fujimoto, Takaaki Hori, Atsushi Nakamura:
Feature space variational Bayesian linear regression and its combination with model space VBLR. ICASSP 2013: 7898-7902 - [c23]Roland Maas, Walter Kellermann, Armin Sehr, Takuya Yoshioka, Marc Delcroix, Keisuke Kinoshita, Tomohiro Nakatani:
Formulation of the REMOS concept from an uncertainty decoding perspective. DSP 2013: 1-6 - [c22]Marc Delcroix, Yotaro Kubo, Tomohiro Nakatani, Atsushi Nakamura:
Is speech enhancement pre-processing still relevant when using deep neural networks for acoustic modeling? INTERSPEECH 2013: 2992-2996 - [c21]Armin Sehr, Takuya Yoshioka, Marc Delcroix, Keisuke Kinoshita, Tomohiro Nakatani, Roland Maas, Walter Kellermann:
Conditional emission densities for combining speech enhancement and recognition systems. INTERSPEECH 2013: 3502-3506 - [c20]Keisuke Kinoshita, Marc Delcroix, Takuya Yoshioka, Tomohiro Nakatani, Armin Sehr, Walter Kellermann, Roland Maas:
The reverb challenge: Acommon evaluation framework for dereverberation and recognition of reverberant speech. WASPAA 2013: 1-4 - 2012
- [j10]Mehrez Souden, Marc Delcroix, Keisuke Kinoshita, Takuya Yoshioka, Tomohiro Nakatani:
Noise Power Spectral Density Tracking: A Maximum Likelihood Perspective. IEEE Signal Process. Lett. 19(8): 495-498 (2012) - [j9]Takuya Yoshioka, Armin Sehr, Marc Delcroix, Keisuke Kinoshita, Roland Maas, Tomohiro Nakatani, Walter Kellermann:
Making Machines Understand Us in Reverberant Rooms: Robustness Against Reverberation for Automatic Speech Recognition. IEEE Signal Process. Mag. 29(6): 114-126 (2012) - [c19]Takuya Yoshioka, Armin Sehr, Marc Delcroix, Keisuke Kinoshita, Roland Maas, Tomohiro Nakatani, Walter Kellermann:
Survey on approaches to speech recognition in reverberant environments. APSIPA 2012: 1-4 - [c18]Tomohiro Nakatani, Takuya Yoshioka, Shoko Araki, Marc Delcroix, Masakiyo Fujimoto:
LogMax observation model with MFCC-based spectral prior for reduction of highly nonstationary ambient noise. ICASSP 2012: 4029-4032 - [c17]Marc Delcroix, Atsunori Ogawa, Shinji Watanabe, Tomohiro Nakatani, Atsushi Nakamura:
Discriminative feature transforms using differenced maximum mutual information. ICASSP 2012: 4753-4756 - [c16]Keisuke Kinoshita, Marc Delcroix, Mehrez Souden, Tomohiro Nakatani:
Example-based speech enhancement with joint utilization of spatial, spectral & temporal cues of speech and noise. INTERSPEECH 2012: 1926-1929 - [c15]Marc Delcroix, Atsunori Ogawa, Tomohiro Nakatani, Atsushi Nakamura:
Dynamic variance adaptation using differenced maximum mutual information. MLSLP 2012: 9-12 - [c14]Mehrez Souden, Keisuke Kinoshita, Marc Delcroix, Tomohiro Nakatani:
Distributed microphone array processing for speech source separation with classifier fusion. MLSP 2012: 1-6 - 2011
- [c13]Keisuke Kinoshita, Mehrez Souden, Marc Delcroix, Tomohiro Nakatani:
Single Channel Dereverberation Using Example-Based Speech Enhancement with Uncertainty Decoding Technique. INTERSPEECH 2011: 197-200 - [c12]Mehrez Souden, Keisuke Kinoshita, Marc Delcroix, Tomohiro Nakatani:
A Multichannel Feature-Based Processing for Robust Speech Recognition. INTERSPEECH 2011: 689-692 - [c11]Tomohiro Nakatani, Shoko Araki, Marc Delcroix, Takuya Yoshioka, Masakiyo Fujimoto:
Reduction of Highly Nonstationary Ambient Noise by Integrating Spectral and Locational Characteristics of Speech and Noise for Robust ASR. INTERSPEECH 2011: 1785-1788 - [p2]Marc Delcroix, Shinji Watanabe, Tomohiro Nakatani:
Variance Compensation for Recognition of Reverberant Speech with Dereverberation Preprocessing. Robust Speech Recognition of Uncertain or Missing Data 2011: 225-255 - 2010
- [p1]Masato Miyoshi, Marc Delcroix, Keisuke Kinoshita, Takuya Yoshioka, Tomohiro Nakatani, Takafumi Hikichi:
Inverse Filtering for Speech Dereverberation Without the Use of Room Acoustics Information. Speech Dereverberation 2010: 271-310
2000 – 2009
- 2009
- [j8]Marc Delcroix, Tomohiro Nakatani, Shinji Watanabe:
Static and Dynamic Variance Compensation for Recognition of Reverberant Speech With Dereverberation Preprocessing. IEEE Trans. Speech Audio Process. 17(2): 324-334 (2009) - [j7]Keisuke Kinoshita, Marc Delcroix, Tomohiro Nakatani, Masato Miyoshi:
Suppression of Late Reverberation Effect on Speech Signal Using Long-Term Multiple-step Linear Prediction. IEEE Trans. Speech Audio Process. 17(4): 534-545 (2009) - 2008
- [j6]Masato Miyoshi, Marc Delcroix, Keisuke Kinoshita:
Calculating Inverse Filters for Speech Dereverberation. IEICE Trans. Fundam. Electron. Commun. Comput. Sci. 91-A(6): 1303-1309 (2008) - [j5]Tomohiro Nakatani, Biing-Hwang Juang, Takuya Yoshioka, Keisuke Kinoshita, Marc Delcroix, Masato Miyoshi:
Speech Dereverberation Based on Maximum-Likelihood Estimation With Time-Varying Gaussian Source Model. IEEE Trans. Speech Audio Process. 16(8): 1512-1527 (2008) - [c10]Marc Delcroix, Tomohiro Nakatani, Shinji Watanabe:
Combined static and dynamic variance adaptation for efficient interconnection of speech enhancement pre-processor with speech recognizer. ICASSP 2008: 4073-4076 - [c9]Dorothea Kolossa, Shoko Araki, Marc Delcroix, Tomohiro Nakatani, Reinhold Orglmeister, Shoji Makino:
Missing feature speech recognition in a meeting situation with maximum SNR beamforming. ISCAS 2008: 3218-3221 - 2007
- [j4]Takafumi Hikichi, Marc Delcroix, Masato Miyoshi:
Inverse Filtering for Speech Dereverberation Less Sensitive to Noise and Room Transfer Function Fluctuations. EURASIP J. Adv. Signal Process. 2007 (2007) - [j3]Marc Delcroix, Takafumi Hikichi, Masato Miyoshi:
Precise Dereverberation Using Multichannel Linear Prediction. IEEE Trans. Speech Audio Process. 15(2): 430-440 (2007) - [j2]Marc Delcroix, Takafumi Hikichi, Masato Miyoshi:
Dereverberation and Denoising Using Multichannel Linear Prediction. IEEE Trans. Speech Audio Process. 15(6): 1791-1801 (2007) - [c8]Tomohiro Nakatani, Biing-Hwang Juang, Takafumi Hikichi, Takuya Yoshioka, Keisuke Kinoshita, Marc Delcroix, Masato Miyoshi:
Study on Speech Dereverberation with Autocorrelation Codebook. ICASSP (1) 2007: 193-196 - [c7]Keisuke Kinoshita, Marc Delcroix, Tomohiro Nakatani, Masato Miyoshi:
Multi-step linear prediction based speech dereverberation in noisy reverberant environment. INTERSPEECH 2007: 854-857 - [c6]Tomohiro Nakatani, Takafumi Hikichi, Keisuke Kinoshita, Takuya Yoshioka, Marc Delcroix, Masato Miyoshi, Biing-Hwang Juang:
Robust blind dereverberation of speech signals based on characteristics of short-time speech segments. ISCAS 2007: 2986-2989 - 2006
- [j1]Marc Delcroix, Takafumi Hikichi, Masato Miyoshi:
On a Blind Speech Dereverberation Algorithm Using Multi-Channel Linear Prediction. IEICE Trans. Fundam. Electron. Commun. Comput. Sci. 89-A(10): 2837-2846 (2006) - [c5]Takafumi Hikichi, Marc Delcroix, Masato Miyoshi:
On robust inverse filter design for room transfer function fluctuations. EUSIPCO 2006: 1-5 - [c4]Marc Delcroix, Takafumi Hikichi, Masato Miyoshi:
On the Use of Lime Dereverberation Algorithm in an Acoustic Environment With a Noise Source. ICASSP (1) 2006: 825-828 - 2005
- [c3]Takafumi Hikichi, Marc Delcroix, Masato Miyoshi:
Blind Dereverberation based on Estimates of Signal Transmission Channels without Precise Information of Channel Order. ICASSP (1) 2005: 1069-1072 - [c2]Marc Delcroix, Takafumi Hikichi, Masato Miyoshi:
Improved blind dereverberation performance by using spatial information. INTERSPEECH 2005: 2309-2312 - 2004
- [c1]Marc Delcroix, Takafumi Hikichi, Masato Miyoshi:
Dereverberation of speech signals based on linear prediction. INTERSPEECH 2004: 877-880
Coauthor Index
aka: Jan Honza Cernocký
aka: Reinhold Haeb-Umbach
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.
Unpaywalled article links
Add open access links from to the list of external document links (if available).
Privacy notice: By enabling the option above, your browser will contact the API of unpaywall.org to load hyperlinks to open access articles. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Unpaywall privacy policy.
Archived links via Wayback Machine
For web page which are no longer available, try to retrieve content from the of the Internet Archive (if available).
Privacy notice: By enabling the option above, your browser will contact the API of archive.org to check for archived content of web pages that are no longer available. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Internet Archive privacy policy.
Reference lists
Add a list of references from , , and to record detail pages.
load references from crossref.org and opencitations.net
Privacy notice: By enabling the option above, your browser will contact the APIs of crossref.org, opencitations.net, and semanticscholar.org to load article reference information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Crossref privacy policy and the OpenCitations privacy policy, as well as the AI2 Privacy Policy covering Semantic Scholar.
Citation data
Add a list of citing articles from and to record detail pages.
load citations from opencitations.net
Privacy notice: By enabling the option above, your browser will contact the API of opencitations.net and semanticscholar.org to load citation information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the OpenCitations privacy policy as well as the AI2 Privacy Policy covering Semantic Scholar.
OpenAlex data
Load additional information about publications from .
Privacy notice: By enabling the option above, your browser will contact the API of openalex.org to load additional information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the information given by OpenAlex.
last updated on 2024-12-01 00:13 CET by the dblp team
all metadata released as open data under CC0 1.0 license
see also: Terms of Use | Privacy Policy | Imprint