[{"status":"public","abstract":[{"lang":"eng","text":"Soft-feature based speech recognition, which is an example of uncertainty decoding, has been proven to be a robust error mitigation method for distributed speech recognition over wireless channels exhibiting bit errors. In this paper we extend this concept to packet-oriented transmissions. The a posteriori probability density function of the lost feature vector, given the closest received neighbours, is computed. In the experiments, the nearest frame repetition, which is shown to be equivalent to the MAP estimate, outperforms the MMSE estimate for long bursts. Taking the variance into account at the speech recognition stage results in superior performance compared to classical schemes using point estimates. A computationally and memory efficient implementation of the proposed packet loss compensation scheme based on table lookup is presented"}],"publication":"IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP 2006)","type":"conference","language":[{"iso":"eng"}],"keyword":["distributed speech recognition","least mean squares methods","MAP estimate","maximum likelihood estimation","MMSE estimate","packet loss compensation scheme","packet switched communication","posteriori probability density function","robust error mitigation method","soft-features","speech recognition","table lookup","voice communication","wireless channels"],"department":[{"_id":"54"}],"user_id":"44006","_id":"11824","intvolume":"         1","page":"I","citation":{"apa":"Ion, V., &#38; Haeb-Umbach, R. (2006). An Inexpensive Packet Loss Compensation Scheme for Distributed Speech Recognition Based on Soft-Features. In <i>IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP 2006)</i> (Vol. 1, p. I). <a href=\"https://doi.org/10.1109/ICASSP.2006.1659984\">https://doi.org/10.1109/ICASSP.2006.1659984</a>","mla":"Ion, Valentin, and Reinhold Haeb-Umbach. “An Inexpensive Packet Loss Compensation Scheme for Distributed Speech Recognition Based on Soft-Features.” <i>IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP 2006)</i>, vol. 1, 2006, p. I, doi:<a href=\"https://doi.org/10.1109/ICASSP.2006.1659984\">10.1109/ICASSP.2006.1659984</a>.","bibtex":"@inproceedings{Ion_Haeb-Umbach_2006, title={An Inexpensive Packet Loss Compensation Scheme for Distributed Speech Recognition Based on Soft-Features}, volume={1}, DOI={<a href=\"https://doi.org/10.1109/ICASSP.2006.1659984\">10.1109/ICASSP.2006.1659984</a>}, booktitle={IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP 2006)}, author={Ion, Valentin and Haeb-Umbach, Reinhold}, year={2006}, pages={I} }","short":"V. Ion, R. Haeb-Umbach, in: IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP 2006), 2006, p. I.","ama":"Ion V, Haeb-Umbach R. An Inexpensive Packet Loss Compensation Scheme for Distributed Speech Recognition Based on Soft-Features. In: <i>IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP 2006)</i>. Vol 1. ; 2006:I. doi:<a href=\"https://doi.org/10.1109/ICASSP.2006.1659984\">10.1109/ICASSP.2006.1659984</a>","ieee":"V. Ion and R. Haeb-Umbach, “An Inexpensive Packet Loss Compensation Scheme for Distributed Speech Recognition Based on Soft-Features,” in <i>IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP 2006)</i>, 2006, vol. 1, p. I.","chicago":"Ion, Valentin, and Reinhold Haeb-Umbach. “An Inexpensive Packet Loss Compensation Scheme for Distributed Speech Recognition Based on Soft-Features.” In <i>IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP 2006)</i>, 1:I, 2006. <a href=\"https://doi.org/10.1109/ICASSP.2006.1659984\">https://doi.org/10.1109/ICASSP.2006.1659984</a>."},"year":"2006","doi":"10.1109/ICASSP.2006.1659984","main_file_link":[{"open_access":"1","url":"https://groups.uni-paderborn.de/nt/pubs/2006/IoHa06-2.pdf"}],"title":"An Inexpensive Packet Loss Compensation Scheme for Distributed Speech Recognition Based on Soft-Features","volume":1,"author":[{"first_name":"Valentin","last_name":"Ion","full_name":"Ion, Valentin"},{"first_name":"Reinhold","full_name":"Haeb-Umbach, Reinhold","id":"242","last_name":"Haeb-Umbach"}],"date_created":"2019-07-12T05:28:58Z","oa":"1","date_updated":"2022-01-06T06:51:10Z"},{"citation":{"ama":"Ion V, Haeb-Umbach R. Uncertainty decoding for distributed speech recognition over error-prone networks. <i>Speech Communication</i>. 2006;48(11):1435-1446. doi:<a href=\"https://doi.org/10.1016/j.specom.2006.03.007\">10.1016/j.specom.2006.03.007</a>","ieee":"V. Ion and R. Haeb-Umbach, “Uncertainty decoding for distributed speech recognition over error-prone networks,” <i>Speech Communication</i>, vol. 48, no. 11, pp. 1435–1446, 2006.","chicago":"Ion, Valentin, and Reinhold Haeb-Umbach. “Uncertainty Decoding for Distributed Speech Recognition over Error-Prone Networks.” <i>Speech Communication</i> 48, no. 11 (2006): 1435–46. <a href=\"https://doi.org/10.1016/j.specom.2006.03.007\">https://doi.org/10.1016/j.specom.2006.03.007</a>.","bibtex":"@article{Ion_Haeb-Umbach_2006, title={Uncertainty decoding for distributed speech recognition over error-prone networks}, volume={48}, DOI={<a href=\"https://doi.org/10.1016/j.specom.2006.03.007\">10.1016/j.specom.2006.03.007</a>}, number={11}, journal={Speech Communication}, author={Ion, Valentin and Haeb-Umbach, Reinhold}, year={2006}, pages={1435–1446} }","mla":"Ion, Valentin, and Reinhold Haeb-Umbach. “Uncertainty Decoding for Distributed Speech Recognition over Error-Prone Networks.” <i>Speech Communication</i>, vol. 48, no. 11, 2006, pp. 1435–46, doi:<a href=\"https://doi.org/10.1016/j.specom.2006.03.007\">10.1016/j.specom.2006.03.007</a>.","short":"V. Ion, R. Haeb-Umbach, Speech Communication 48 (2006) 1435–1446.","apa":"Ion, V., &#38; Haeb-Umbach, R. (2006). Uncertainty decoding for distributed speech recognition over error-prone networks. <i>Speech Communication</i>, <i>48</i>(11), 1435–1446. <a href=\"https://doi.org/10.1016/j.specom.2006.03.007\">https://doi.org/10.1016/j.specom.2006.03.007</a>"},"intvolume":"        48","page":"1435-1446","main_file_link":[{"open_access":"1","url":"https://groups.uni-paderborn.de/nt/pubs/2006/IoHa06-3.pdf"}],"doi":"10.1016/j.specom.2006.03.007","author":[{"first_name":"Valentin","full_name":"Ion, Valentin","last_name":"Ion"},{"first_name":"Reinhold","full_name":"Haeb-Umbach, Reinhold","id":"242","last_name":"Haeb-Umbach"}],"volume":48,"oa":"1","date_updated":"2022-01-06T06:51:10Z","status":"public","type":"journal_article","user_id":"44006","department":[{"_id":"54"}],"_id":"11825","year":"2006","issue":"11","title":"Uncertainty decoding for distributed speech recognition over error-prone networks","date_created":"2019-07-12T05:28:59Z","abstract":[{"lang":"eng","text":"In this paper, we propose an enhanced error concealment strategy at the server side of a distributed speech recognition (DSR) system, which is fully compatible with the existing DSR standard. It is based on a Bayesian approach, where the a posteriori probability density of the error-free feature vector is computed, given all received feature vectors which are possibly corrupted by transmission errors. Rather than computing a point estimate, such as the MMSE estimate, and plugging it into the Bayesian decision rule, we employ uncertainty decoding, which results in an integration over the uncertainty in the feature domain. In a typical scenario the communication between the thin client, often a mobile device, and the recognition server spreads across heterogeneous networks. Both bit errors on circuit-switched links and lost data packets on IP connections are mitigated by our approach in a unified manner. The experiments reveal improved robustness both for small- and large-vocabulary recognition tasks."}],"publication":"Speech Communication","language":[{"iso":"eng"}],"keyword":["Channel error robustness","Distributed speech recognition","Soft features","Uncertainty decoding"]},{"type":"conference","publication":"IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP 2006)","abstract":[{"text":"A marginalized particle filter is proposed for performing single channel speech enhancement with a non-linear dynamic state model. The system consists of a particle filter for tracking line spectral pair (LSP) parameters and a Kalman filter per particle for speech enhancement. The state model for the LSPs has been learnt on clean speech training data. In our approach parameters and speech samples are processed at different time scales by assuming the parameters to be constant for small blocks of data. Further enhancement is obtained by an iteration which can be applied on these small blocks. The experiments show that similar SNR gains are obtained as with the Kalman-LM-iterative algorithm. However better values of the noise level and the log-spectral distance are achieved","lang":"eng"}],"status":"public","_id":"11943","user_id":"44006","department":[{"_id":"54"}],"keyword":["clean speech training data","iterative methods","iterative speech enhancement","Kalman filter","Kalman filters","Kalman-LM-iterative algorithm","line spectral pair parameters","log-spectral distance","marginalized particle filter","noise level","nonlinear dynamic state speech model","particle filtering (numerical methods)","single channel speech enhancement","SNR gains","speech enhancement","speech samples"],"language":[{"iso":"eng"}],"year":"2006","citation":{"ieee":"S. Windmann and R. Haeb-Umbach, “Iterative Speech Enhancement using a Non-Linear Dynamic State Model of Speech and its Parameters,” in <i>IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP 2006)</i>, 2006, vol. 1, p. I.","chicago":"Windmann, Stefan, and Reinhold Haeb-Umbach. “Iterative Speech Enhancement Using a Non-Linear Dynamic State Model of Speech and Its Parameters.” In <i>IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP 2006)</i>, 1:I, 2006. <a href=\"https://doi.org/10.1109/ICASSP.2006.1660058\">https://doi.org/10.1109/ICASSP.2006.1660058</a>.","ama":"Windmann S, Haeb-Umbach R. Iterative Speech Enhancement using a Non-Linear Dynamic State Model of Speech and its Parameters. In: <i>IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP 2006)</i>. Vol 1. ; 2006:I. doi:<a href=\"https://doi.org/10.1109/ICASSP.2006.1660058\">10.1109/ICASSP.2006.1660058</a>","bibtex":"@inproceedings{Windmann_Haeb-Umbach_2006, title={Iterative Speech Enhancement using a Non-Linear Dynamic State Model of Speech and its Parameters}, volume={1}, DOI={<a href=\"https://doi.org/10.1109/ICASSP.2006.1660058\">10.1109/ICASSP.2006.1660058</a>}, booktitle={IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP 2006)}, author={Windmann, Stefan and Haeb-Umbach, Reinhold}, year={2006}, pages={I} }","mla":"Windmann, Stefan, and Reinhold Haeb-Umbach. “Iterative Speech Enhancement Using a Non-Linear Dynamic State Model of Speech and Its Parameters.” <i>IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP 2006)</i>, vol. 1, 2006, p. I, doi:<a href=\"https://doi.org/10.1109/ICASSP.2006.1660058\">10.1109/ICASSP.2006.1660058</a>.","short":"S. Windmann, R. Haeb-Umbach, in: IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP 2006), 2006, p. I.","apa":"Windmann, S., &#38; Haeb-Umbach, R. (2006). Iterative Speech Enhancement using a Non-Linear Dynamic State Model of Speech and its Parameters. In <i>IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP 2006)</i> (Vol. 1, p. I). <a href=\"https://doi.org/10.1109/ICASSP.2006.1660058\">https://doi.org/10.1109/ICASSP.2006.1660058</a>"},"intvolume":"         1","page":"I","oa":"1","date_updated":"2022-01-06T06:51:12Z","date_created":"2019-07-12T05:31:15Z","author":[{"first_name":"Stefan","full_name":"Windmann, Stefan","last_name":"Windmann"},{"id":"242","full_name":"Haeb-Umbach, Reinhold","last_name":"Haeb-Umbach","first_name":"Reinhold"}],"volume":1,"title":"Iterative Speech Enhancement using a Non-Linear Dynamic State Model of Speech and its Parameters","main_file_link":[{"open_access":"1","url":"https://groups.uni-paderborn.de/nt/pubs/2006/WiHa06-2.pdf"}],"doi":"10.1109/ICASSP.2006.1660058"},{"citation":{"ama":"Ion V, Haeb-Umbach R. A Comparison of Soft-Feature Distributed Speech Recognition with Candidate Codecs for Speech Enabled Mobile Services. In: <i>IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP 2005)</i>. Vol 1. ; 2005:333-336. doi:<a href=\"https://doi.org/10.1109/ICASSP.2005.1415118\">10.1109/ICASSP.2005.1415118</a>","ieee":"V. Ion and R. Haeb-Umbach, “A Comparison of Soft-Feature Distributed Speech Recognition with Candidate Codecs for Speech Enabled Mobile Services,” in <i>IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP 2005)</i>, 2005, vol. 1, pp. 333–336.","chicago":"Ion, Valentin, and Reinhold Haeb-Umbach. “A Comparison of Soft-Feature Distributed Speech Recognition with Candidate Codecs for Speech Enabled Mobile Services.” In <i>IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP 2005)</i>, 1:333–36, 2005. <a href=\"https://doi.org/10.1109/ICASSP.2005.1415118\">https://doi.org/10.1109/ICASSP.2005.1415118</a>.","bibtex":"@inproceedings{Ion_Haeb-Umbach_2005, title={A Comparison of Soft-Feature Distributed Speech Recognition with Candidate Codecs for Speech Enabled Mobile Services}, volume={1}, DOI={<a href=\"https://doi.org/10.1109/ICASSP.2005.1415118\">10.1109/ICASSP.2005.1415118</a>}, booktitle={IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP 2005)}, author={Ion, Valentin and Haeb-Umbach, Reinhold}, year={2005}, pages={333–336} }","short":"V. Ion, R. Haeb-Umbach, in: IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP 2005), 2005, pp. 333–336.","mla":"Ion, Valentin, and Reinhold Haeb-Umbach. “A Comparison of Soft-Feature Distributed Speech Recognition with Candidate Codecs for Speech Enabled Mobile Services.” <i>IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP 2005)</i>, vol. 1, 2005, pp. 333–36, doi:<a href=\"https://doi.org/10.1109/ICASSP.2005.1415118\">10.1109/ICASSP.2005.1415118</a>.","apa":"Ion, V., &#38; Haeb-Umbach, R. (2005). A Comparison of Soft-Feature Distributed Speech Recognition with Candidate Codecs for Speech Enabled Mobile Services. In <i>IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP 2005)</i> (Vol. 1, pp. 333–336). <a href=\"https://doi.org/10.1109/ICASSP.2005.1415118\">https://doi.org/10.1109/ICASSP.2005.1415118</a>"},"page":"333-336","intvolume":"         1","year":"2005","author":[{"first_name":"Valentin","last_name":"Ion","full_name":"Ion, Valentin"},{"full_name":"Haeb-Umbach, Reinhold","id":"242","last_name":"Haeb-Umbach","first_name":"Reinhold"}],"date_created":"2019-07-12T05:29:02Z","volume":1,"date_updated":"2022-01-06T06:51:10Z","oa":"1","main_file_link":[{"url":"https://groups.uni-paderborn.de/nt/pubs/2005/IoHa05-2.pdf","open_access":"1"}],"doi":"10.1109/ICASSP.2005.1415118","title":"A Comparison of Soft-Feature Distributed Speech Recognition with Candidate Codecs for Speech Enabled Mobile Services","type":"conference","publication":"IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP 2005)","status":"public","abstract":[{"text":"In this paper we present a comparison of the recently proposed Soft-Feature Distributed Speech Recognition (SFDSR) with the two evaluated candidate codecs for Speech Enabled Services over wireless networks: Adaptive Multirate Codec (AMR) and the ETSI Extended Advanced Front-End for Distributed Speech Recognition (XAFE). It is shown that SFDSR achieves the best recognition performance on a simulated GSM transmission, followed by XAFE and AMR.We also present some new results concerning SFDSR which demonstrate the versatility of the approach. Further, a simple method is introduced which considerably reduces the computational effort.","lang":"eng"}],"user_id":"44006","department":[{"_id":"54"}],"_id":"11828","language":[{"iso":"eng"}],"keyword":["adaptive codes","adaptive multirate codec","AMR","distributed speech recognition","ETSI","extended advanced front-end","recognition performance","SFDSR","simulated GSM transmission","soft-feature distributed speech recognition","speech codecs","speech coding","speech recognition","variable rate codes","XAFE"]},{"publication":"IEEE Workshop on Multimedia Signal Processing (MMSP 2004)","type":"conference","abstract":[{"text":"The paper is concerned with binaural signal processing for a bimodal human-robot interface with hearing and vision. The two microphone signals are processed to obtain an enhanced single-channel input signal for the subsequent speech recognizer and to localize the acoustic source, an important information for establishing a natural human-robot communication. We utilize a robust adaptive algorithm for filter-and-sum beamforming (FSB) and extract speaker direction information from the resulting FIR filter coefficients. Further, particle filtering is applied which conducts a nonlinear Bayesian tracking of speaker movement. Good location accuracy can be achieved even in highly reverberant environments. The results obtained outperform the conventional generalized cross correlation (GCC) method.","lang":"eng"}],"status":"public","_id":"11931","department":[{"_id":"54"}],"user_id":"44006","keyword":["bimodal human-robot interface","binaural signal processing","enhanced single-channel input signal","filter-and-sum beamforming","filtering theory","FIR filter coefficient","generalized cross correlation method","microphones","microphone signal","nonlinear Bayesian tracking","particle filtering","robust adaptive algorithm","robust speaker direction estimation","signal processing","speech enhancement","speech recognition","speech recognizer","user interfaces"],"language":[{"iso":"eng"}],"year":"2004","page":"367-370","citation":{"chicago":"Warsitz, Ernst, and Reinhold Haeb-Umbach. “Robust Speaker Direction Estimation with Particle Filtering.” In <i>IEEE Workshop on Multimedia Signal Processing (MMSP 2004)</i>, 367–70, 2004. <a href=\"https://doi.org/10.1109/MMSP.2004.1436569\">https://doi.org/10.1109/MMSP.2004.1436569</a>.","ieee":"E. Warsitz and R. Haeb-Umbach, “Robust speaker direction estimation with particle filtering,” in <i>IEEE Workshop on Multimedia Signal Processing (MMSP 2004)</i>, 2004, pp. 367–370.","ama":"Warsitz E, Haeb-Umbach R. Robust speaker direction estimation with particle filtering. In: <i>IEEE Workshop on Multimedia Signal Processing (MMSP 2004)</i>. ; 2004:367-370. doi:<a href=\"https://doi.org/10.1109/MMSP.2004.1436569\">10.1109/MMSP.2004.1436569</a>","apa":"Warsitz, E., &#38; Haeb-Umbach, R. (2004). Robust speaker direction estimation with particle filtering. In <i>IEEE Workshop on Multimedia Signal Processing (MMSP 2004)</i> (pp. 367–370). <a href=\"https://doi.org/10.1109/MMSP.2004.1436569\">https://doi.org/10.1109/MMSP.2004.1436569</a>","bibtex":"@inproceedings{Warsitz_Haeb-Umbach_2004, title={Robust speaker direction estimation with particle filtering}, DOI={<a href=\"https://doi.org/10.1109/MMSP.2004.1436569\">10.1109/MMSP.2004.1436569</a>}, booktitle={IEEE Workshop on Multimedia Signal Processing (MMSP 2004)}, author={Warsitz, Ernst and Haeb-Umbach, Reinhold}, year={2004}, pages={367–370} }","mla":"Warsitz, Ernst, and Reinhold Haeb-Umbach. “Robust Speaker Direction Estimation with Particle Filtering.” <i>IEEE Workshop on Multimedia Signal Processing (MMSP 2004)</i>, 2004, pp. 367–70, doi:<a href=\"https://doi.org/10.1109/MMSP.2004.1436569\">10.1109/MMSP.2004.1436569</a>.","short":"E. Warsitz, R. Haeb-Umbach, in: IEEE Workshop on Multimedia Signal Processing (MMSP 2004), 2004, pp. 367–370."},"date_updated":"2022-01-06T06:51:12Z","oa":"1","author":[{"first_name":"Ernst","full_name":"Warsitz, Ernst","last_name":"Warsitz"},{"last_name":"Haeb-Umbach","id":"242","full_name":"Haeb-Umbach, Reinhold","first_name":"Reinhold"}],"date_created":"2019-07-12T05:31:01Z","title":"Robust speaker direction estimation with particle filtering","doi":"10.1109/MMSP.2004.1436569","main_file_link":[{"open_access":"1","url":"https://groups.uni-paderborn.de/nt/pubs/2004/WaHa04.pdf"}]},{"department":[{"_id":"672"}],"user_id":"5786","_id":"39053","language":[{"iso":"eng"}],"keyword":["User interfaces","Speech recognition","Streaming media","Specification languages","Keyboards","Speech synthesis","Rendering (computer graphics)","Ambient intelligence","Humans","Displays"],"publication":"Proceedings of HICCS-37","type":"conference","status":"public","abstract":[{"text":"Portable devices come with different limitations in user interaction like limited display size, small keyboard, and different sorts of input and output capabilities. With the advance of speech recognition and speech synthesis technologies, their complementary use becomes attractive for mobile devices in order to implement real multimodal user interaction. However, current systems and formats do not sufficiently integrate advanced multimodal interactions. We introduce an advanced generic multimodal interaction and rendering system (MIRS) dedicated for mobile devices. MIRS incorporates efficient processing of XML specification languages for limited, mobile devices and comes with the XML-based dialog and interface specification language (DISL). DISL can be considered as an UIML subset, which is enhanced by the means of state-oriented dialog specifications. The dialog specification is based on ODSN (object oriented dialog specification notation), which has been introduced to define user interface control by means of interaction states with transition rules.","lang":"eng"}],"date_created":"2023-01-24T08:46:31Z","author":[{"id":"16243","full_name":"Müller, Wolfgang","last_name":"Müller","first_name":"Wolfgang"},{"full_name":"Schäfer, Robbie","last_name":"Schäfer","first_name":"Robbie"},{"last_name":"Bleul","full_name":"Bleul, Steffen","first_name":"Steffen"}],"date_updated":"2023-01-24T08:46:37Z","doi":"10.1109/HICSS.2004.1265674","conference":{"location":"Waikoloa, HI, USA","name":"37th Annual Hawaii International Conference on System Sciences"},"title":"Interactive Multimodal User Interfaces for Mobile Devices","publication_identifier":{"isbn":["0-7695-2056-1"]},"citation":{"chicago":"Müller, Wolfgang, Robbie Schäfer, and Steffen Bleul. “Interactive Multimodal User Interfaces for Mobile Devices.” In <i>Proceedings of HICCS-37</i>. Waikoloa, HI, USA, 2004. <a href=\"https://doi.org/10.1109/HICSS.2004.1265674\">https://doi.org/10.1109/HICSS.2004.1265674</a>.","ieee":"W. Müller, R. Schäfer, and S. Bleul, “Interactive Multimodal User Interfaces for Mobile Devices,” presented at the 37th Annual Hawaii International Conference on System Sciences, Waikoloa, HI, USA, 2004, doi: <a href=\"https://doi.org/10.1109/HICSS.2004.1265674\">10.1109/HICSS.2004.1265674</a>.","ama":"Müller W, Schäfer R, Bleul S. Interactive Multimodal User Interfaces for Mobile Devices. In: <i>Proceedings of HICCS-37</i>. ; 2004. doi:<a href=\"https://doi.org/10.1109/HICSS.2004.1265674\">10.1109/HICSS.2004.1265674</a>","apa":"Müller, W., Schäfer, R., &#38; Bleul, S. (2004). Interactive Multimodal User Interfaces for Mobile Devices. <i>Proceedings of HICCS-37</i>. 37th Annual Hawaii International Conference on System Sciences, Waikoloa, HI, USA. <a href=\"https://doi.org/10.1109/HICSS.2004.1265674\">https://doi.org/10.1109/HICSS.2004.1265674</a>","bibtex":"@inproceedings{Müller_Schäfer_Bleul_2004, place={Waikoloa, HI, USA}, title={Interactive Multimodal User Interfaces for Mobile Devices}, DOI={<a href=\"https://doi.org/10.1109/HICSS.2004.1265674\">10.1109/HICSS.2004.1265674</a>}, booktitle={Proceedings of HICCS-37}, author={Müller, Wolfgang and Schäfer, Robbie and Bleul, Steffen}, year={2004} }","mla":"Müller, Wolfgang, et al. “Interactive Multimodal User Interfaces for Mobile Devices.” <i>Proceedings of HICCS-37</i>, 2004, doi:<a href=\"https://doi.org/10.1109/HICSS.2004.1265674\">10.1109/HICSS.2004.1265674</a>.","short":"W. Müller, R. Schäfer, S. Bleul, in: Proceedings of HICCS-37, Waikoloa, HI, USA, 2004."},"place":"Waikoloa, HI, USA","year":"2004"},{"abstract":[{"text":"In this paper, it is shown that a correlation criterion is the appropriate criterion for bottom-up clustering to obtain broad phonetic class regression trees for maximum likelihood linear regression (MLLR)-based speaker adaptation. The correlation structure among speech units is estimated on the speaker-independent training data. In adaptation experiments the tree outperformed a regression tree obtained from clustering according to closeness in acoustic space and achieved results comparable with those of a manually designed broad phonetic class tree","lang":"eng"}],"status":"public","type":"journal_article","publication":"IEEE Transactions on Speech and Audio Processing","keyword":["acoustic space","adaptation experiments","automatic generation","bottom-up clustering","broad phonetic class regression trees","correlation criterion","correlation methods","maximum likelihood estimation","maximum likelihood linear regression based speaker adaptation","MLLR adaptation","pattern clustering","phonetic regression class trees","speaker-independent training data","speech recognition","speech units","statistical analysis","trees (mathematics)"],"language":[{"iso":"eng"}],"_id":"11778","user_id":"44006","department":[{"_id":"54"}],"year":"2001","citation":{"ama":"Haeb-Umbach R. Automatic generation of phonetic regression class trees for MLLR adaptation. <i>IEEE Transactions on Speech and Audio Processing</i>. 2001;9(3):299-302. doi:<a href=\"https://doi.org/10.1109/89.906003\">10.1109/89.906003</a>","chicago":"Haeb-Umbach, Reinhold. “Automatic Generation of Phonetic Regression Class Trees for MLLR Adaptation.” <i>IEEE Transactions on Speech and Audio Processing</i> 9, no. 3 (2001): 299–302. <a href=\"https://doi.org/10.1109/89.906003\">https://doi.org/10.1109/89.906003</a>.","ieee":"R. Haeb-Umbach, “Automatic generation of phonetic regression class trees for MLLR adaptation,” <i>IEEE Transactions on Speech and Audio Processing</i>, vol. 9, no. 3, pp. 299–302, 2001.","bibtex":"@article{Haeb-Umbach_2001, title={Automatic generation of phonetic regression class trees for MLLR adaptation}, volume={9}, DOI={<a href=\"https://doi.org/10.1109/89.906003\">10.1109/89.906003</a>}, number={3}, journal={IEEE Transactions on Speech and Audio Processing}, author={Haeb-Umbach, Reinhold}, year={2001}, pages={299–302} }","short":"R. Haeb-Umbach, IEEE Transactions on Speech and Audio Processing 9 (2001) 299–302.","mla":"Haeb-Umbach, Reinhold. “Automatic Generation of Phonetic Regression Class Trees for MLLR Adaptation.” <i>IEEE Transactions on Speech and Audio Processing</i>, vol. 9, no. 3, 2001, pp. 299–302, doi:<a href=\"https://doi.org/10.1109/89.906003\">10.1109/89.906003</a>.","apa":"Haeb-Umbach, R. (2001). Automatic generation of phonetic regression class trees for MLLR adaptation. <i>IEEE Transactions on Speech and Audio Processing</i>, <i>9</i>(3), 299–302. <a href=\"https://doi.org/10.1109/89.906003\">https://doi.org/10.1109/89.906003</a>"},"intvolume":"         9","page":"299-302","issue":"3","title":"Automatic generation of phonetic regression class trees for MLLR adaptation","main_file_link":[{"url":"https://groups.uni-paderborn.de/nt/pubs/2001/Ha01.pdf","open_access":"1"}],"doi":"10.1109/89.906003","date_updated":"2022-01-06T06:51:08Z","oa":"1","author":[{"id":"242","full_name":"Haeb-Umbach, Reinhold","last_name":"Haeb-Umbach","first_name":"Reinhold"}],"date_created":"2019-07-12T05:28:04Z","volume":9},{"user_id":"24135","department":[{"_id":"518"}],"_id":"2433","keyword":["co-design","speech processing"],"type":"mastersthesis","status":"public","author":[{"id":"16153","full_name":"Plessl, Christian","last_name":"Plessl","orcid":"0000-0001-5728-9982","first_name":"Christian"},{"first_name":"Simon","full_name":"Maurer, Simon","last_name":"Maurer"}],"date_created":"2018-04-17T15:56:00Z","date_updated":"2022-01-06T06:56:17Z","publisher":"Computer Engineering and Networks Lab, ETH Zurich, Switzerland","title":"Hardware/Software Codesign in Speech Compression Applications","citation":{"ama":"Plessl C, Maurer S. <i>Hardware/Software Codesign in Speech Compression Applications</i>. Computer Engineering and Networks Lab, ETH Zurich, Switzerland; 2000.","ieee":"C. Plessl and S. Maurer, <i>Hardware/Software Codesign in Speech Compression Applications</i>. Computer Engineering and Networks Lab, ETH Zurich, Switzerland, 2000.","chicago":"Plessl, Christian, and Simon Maurer. <i>Hardware/Software Codesign in Speech Compression Applications</i>. Computer Engineering and Networks Lab, ETH Zurich, Switzerland, 2000.","mla":"Plessl, Christian, and Simon Maurer. <i>Hardware/Software Codesign in Speech Compression Applications</i>. Computer Engineering and Networks Lab, ETH Zurich, Switzerland, 2000.","short":"C. Plessl, S. Maurer, Hardware/Software Codesign in Speech Compression Applications, Computer Engineering and Networks Lab, ETH Zurich, Switzerland, 2000.","bibtex":"@book{Plessl_Maurer_2000, title={Hardware/Software Codesign in Speech Compression Applications}, publisher={Computer Engineering and Networks Lab, ETH Zurich, Switzerland}, author={Plessl, Christian and Maurer, Simon}, year={2000} }","apa":"Plessl, C., &#38; Maurer, S. (2000). <i>Hardware/Software Codesign in Speech Compression Applications</i>. Computer Engineering and Networks Lab, ETH Zurich, Switzerland."},"year":"2000"},{"date_created":"2019-07-12T05:29:50Z","author":[{"last_name":"Lieb","full_name":"Lieb, M.","first_name":"M."},{"full_name":"Haeb-Umbach, Reinhold","id":"242","last_name":"Haeb-Umbach","first_name":"Reinhold"}],"volume":2,"date_updated":"2022-01-06T06:51:11Z","oa":"1","main_file_link":[{"url":"https://groups.uni-paderborn.de/nt/pubs/2000/LiHa00.pdf","open_access":"1"}],"doi":"10.1109/ICASSP.2000.859157","title":"LDA derived cepstral trajectory filters in adverse environmental conditions","citation":{"ama":"Lieb M, Haeb-Umbach R. LDA derived cepstral trajectory filters in adverse environmental conditions. In: <i>IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP 2000)</i>. Vol 2. ; 2000:II1105-II1108 vol.2. doi:<a href=\"https://doi.org/10.1109/ICASSP.2000.859157\">10.1109/ICASSP.2000.859157</a>","chicago":"Lieb, M., and Reinhold Haeb-Umbach. “LDA Derived Cepstral Trajectory Filters in Adverse Environmental Conditions.” In <i>IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP 2000)</i>, 2:II1105-II1108 vol.2, 2000. <a href=\"https://doi.org/10.1109/ICASSP.2000.859157\">https://doi.org/10.1109/ICASSP.2000.859157</a>.","ieee":"M. Lieb and R. Haeb-Umbach, “LDA derived cepstral trajectory filters in adverse environmental conditions,” in <i>IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP 2000)</i>, 2000, vol. 2, pp. II1105-II1108 vol.2.","apa":"Lieb, M., &#38; Haeb-Umbach, R. (2000). LDA derived cepstral trajectory filters in adverse environmental conditions. In <i>IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP 2000)</i> (Vol. 2, pp. II1105-II1108 vol.2). <a href=\"https://doi.org/10.1109/ICASSP.2000.859157\">https://doi.org/10.1109/ICASSP.2000.859157</a>","bibtex":"@inproceedings{Lieb_Haeb-Umbach_2000, title={LDA derived cepstral trajectory filters in adverse environmental conditions}, volume={2}, DOI={<a href=\"https://doi.org/10.1109/ICASSP.2000.859157\">10.1109/ICASSP.2000.859157</a>}, booktitle={IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP 2000)}, author={Lieb, M. and Haeb-Umbach, Reinhold}, year={2000}, pages={II1105-II1108 vol.2} }","short":"M. Lieb, R. Haeb-Umbach, in: IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP 2000), 2000, pp. II1105-II1108 vol.2.","mla":"Lieb, M., and Reinhold Haeb-Umbach. “LDA Derived Cepstral Trajectory Filters in Adverse Environmental Conditions.” <i>IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP 2000)</i>, vol. 2, 2000, pp. II1105-II1108 vol.2, doi:<a href=\"https://doi.org/10.1109/ICASSP.2000.859157\">10.1109/ICASSP.2000.859157</a>."},"page":"II1105-II1108 vol.2","intvolume":"         2","year":"2000","user_id":"44006","department":[{"_id":"54"}],"_id":"11869","language":[{"iso":"eng"}],"keyword":["acoustic echo cancellation algorithms","adverse environmental conditions","automatic speech recognition","cepstral analysis","cepstral features","cepstral mean normalization","command word task","delta-delta features","delta features","echo suppression","error rate reductions","feature vector components","FIR filters","LDA derived cepstral trajectory filters","linear discriminant analysis","long-range feature filters","phone accuracy","real-life room impulse responses","reverberant data","spectral parameters","speech recognition","standard TIMIT phone recognition task"],"type":"conference","publication":"IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP 2000)","status":"public","abstract":[{"text":"Amongst several data driven approaches for designing filters for the time sequence of spectral parameters, the linear discriminant analysis (LDA) based method has been proposed for automatic speech recognition. Here we apply LDA-based filter design to cepstral features, which better match the inherent assumption of this method that feature vector components are uncorrelated. Extensive recognition experiments have been conducted both on the standard TIMIT phone recognition task and on a proprietary 130-words command word task under various adverse environmental conditions, including reverberant data with real-life room impulse responses and data processed by acoustic echo cancellation algorithms. Significant error rate reductions have been achieved when applying the novel long-range feature filters compared to standard approaches employing cepstral mean normalization and delta and delta-delta features, in particular when facing acoustic echo cancellation scenarios and room reverberation. For example, the phone accuracy on reverberated TIMIT data could be increased from 50.7\\% to 56.0\\%","lang":"eng"}]}]
