@inproceedings{14822, abstract = {{Multi-talker speech and moving speakers still pose a significant challenge to automatic speech recognition systems. Assuming an enrollment utterance of the target speakeris available, the so-called SpeakerBeam concept has been recently proposed to extract the target speaker from a speech mixture. If multi-channel input is available, spatial properties of the speaker can be exploited to support the source extraction. In this contribution we investigate different approaches to exploit such spatial information. In particular, we are interested in the question, how useful this information is if the target speaker changes his/her position. To this end, we present a SpeakerBeam-based source extraction network that is adapted to work on moving speakers by recursively updating the beamformer coefficients. Experimental results are presented on two data sets, one with articially created room impulse responses, and one with real room impulse responses and noise recorded in a conference room. Interestingly, spatial features turn out to be advantageous even if the speaker position changes.}}, author = {{Heitkaemper, Jens and Feher, Thomas and Freitag, Michael and Haeb-Umbach, Reinhold}}, booktitle = {{International Conference on Statistical Language and Speech Processing 2019, Ljubljana, Slovenia}}, title = {{{A Study on Online Source Extraction in the Presence of Changing Speaker Positions}}}, year = {{2019}}, } @inproceedings{14824, abstract = {{This paper deals with multi-channel speech recognition in scenarios with multiple speakers. Recently, the spectral characteristics of a target speaker, extracted from an adaptation utterance, have been used to guide a neural network mask estimator to focus on that speaker. In this work we present two variants of speakeraware neural networks, which exploit both spectral and spatial information to allow better discrimination between target and interfering speakers. Thus, we introduce either a spatial preprocessing prior to the mask estimation or a spatial plus spectral speaker characterization block whose output is directly fed into the neural mask estimator. The target speaker’s spectral and spatial signature is extracted from an adaptation utterance recorded at the beginning of a session. We further adapt the architecture for low-latency processing by means of block-online beamforming that recursively updates the signal statistics. Experimental results show that the additional spatial information clearly improves source extraction, in particular in the same-gender case, and that our proposal achieves state-of-the-art performance in terms of distortion reduction and recognition accuracy.}}, author = {{Martin-Donas, Juan M. and Heitkaemper, Jens and Haeb-Umbach, Reinhold and Gomez, Angel M. and Peinado, Antonio M.}}, booktitle = {{INTERSPEECH 2019, Graz, Austria}}, title = {{{Multi-Channel Block-Online Source Extraction based on Utterance Adaptation}}}, year = {{2019}}, } @inproceedings{14826, abstract = {{In this paper, we present Hitachi and Paderborn University’s joint effort for automatic speech recognition (ASR) in a dinner party scenario. The main challenges of ASR systems for dinner party recordings obtained by multiple microphone arrays are (1) heavy speech overlaps, (2) severe noise and reverberation, (3) very natural onversational content, and possibly (4) insufficient training data. As an example of a dinner party scenario, we have chosen the data presented during the CHiME-5 speech recognition challenge, where the baseline ASR had a 73.3% word error rate (WER), and even the best performing system at the CHiME-5 challenge had a 46.1% WER. We extensively investigated a combination of the guided source separation-based speech enhancement technique and an already proposed strong ASR backend and found that a tight combination of these techniques provided substantial accuracy improvements. Our final system achieved WERs of 39.94% and 41.64% for the development and evaluation data, respectively, both of which are the best published results for the dataset. We also investigated with additional training data on the official small data in the CHiME-5 corpus to assess the intrinsic difficulty of this ASR task.}}, author = {{Kanda, Naoyuki and Boeddeker, Christoph and Heitkaemper, Jens and Fujita, Yusuke and Horiguchi, Shota and Haeb-Umbach, Reinhold}}, booktitle = {{INTERSPEECH 2019, Graz, Austria}}, title = {{{Guided Source Separation Meets a Strong ASR Backend: Hitachi/Paderborn University Joint Investigation for Dinner Party ASR}}}, year = {{2019}}, } @techreport{14902, author = {{Mair, Christina and Scheffler, Wolfram and Senger, Isabell and Sureth-Sloane, Caren}}, title = {{{Analyse der Veränderung der zwischenstaatlichen Gewinnaufteilung bei Einführung einer standardisierten Gewinnverteilungsmethode am Beispiel des Einsatzes von 3D-Druckern}}}, volume = {{42}}, year = {{2019}}, } @article{14990, abstract = {{We investigate optical microresonators consisting of either one or two coupled rectangular strips between upper and lower slab waveguides. The cavities are evanescently excited under oblique angles by thin-film guided, in-plane unguided waves supported by one of the slab waveguides. Beyond a specific incidence angle, losses are fully suppressed. The interaction between the guided mode of the cavity-strip and the incoming slab modes leads to resonant behavior for specific incidence angles and gaps. For a single cavity, at resonance, the input power is equally split among each of the four output ports, while for two cavities an add-drop filter can be realized that, at resonance, routes the incoming power completely to the forward drop waveguide via the cavity. For both applications, the strength of the interaction is controlled by the gaps between cavities and waveguides.}}, author = {{Ebers, Lena and Hammer, Manfred and Berkemeier, Manuel B. and Menzel, Alexander and Förstner, Jens}}, issn = {{2578-7519}}, journal = {{OSA Continuum}}, keywords = {{tet_topic_waveguides}}, pages = {{3288}}, title = {{{Coupled microstrip-cavities under oblique incidence of semi-guided waves: a lossless integrated optical add-drop filter}}}, doi = {{10.1364/osac.2.003288}}, volume = {{2}}, year = {{2019}}, } @article{15002, abstract = {{Many problem settings in machine learning are concerned with the simultaneous prediction of multiple target variables of diverse type. Amongst others, such problem settings arise in multivariate regression, multi-label classification, multi-task learning, dyadic prediction, zero-shot learning, network inference, and matrix completion. These subfields of machine learning are typically studied in isolation, without highlighting or exploring important relationships. In this paper, we present a unifying view on what we call multi-target prediction (MTP) problems and methods. First, we formally discuss commonalities and differences between existing MTP problems. To this end, we introduce a general framework that covers the above subfields as special cases. As a second contribution, we provide a structured overview of MTP methods. This is accomplished by identifying a number of key properties, which distinguish such methods and determine their suitability for different types of problems. Finally, we also discuss a few challenges for future research.}}, author = {{Waegeman, Willem and Dembczynski, Krzysztof and Hüllermeier, Eyke}}, issn = {{1573-756X}}, journal = {{Data Mining and Knowledge Discovery}}, number = {{2}}, pages = {{293--324}}, title = {{{Multi-target prediction: a unifying view on problems and methods}}}, doi = {{10.1007/s10618-018-0595-5}}, volume = {{33}}, year = {{2019}}, } @inproceedings{15007, author = {{Melnikov, Vitaly and Hüllermeier, Eyke}}, booktitle = {{Proceedings ACML, Asian Conference on Machine Learning (Proceedings of Machine Learning Research, 101)}}, title = {{{Learning to Aggregate: Tackling the Aggregation/Disaggregation Problem for OWA}}}, doi = {{10.1016/j.jmva.2019.02.017}}, year = {{2019}}, } @inproceedings{15011, author = {{Tornede, Alexander and Wever, Marcel Dominik and Hüllermeier, Eyke}}, booktitle = {{Proceedings - 29. Workshop Computational Intelligence, Dortmund, 28. - 29. November 2019}}, editor = {{Hoffmann, Frank and Hüllermeier, Eyke and Mikut, Ralf}}, isbn = {{978-3-7315-0979-0}}, location = {{Dortmund}}, pages = {{135--146}}, publisher = {{KIT Scientific Publishing, Karlsruhe}}, title = {{{Algorithm Selection as Recommendation: From Collaborative Filtering to Dyad Ranking}}}, year = {{2019}}, } @article{16337, author = {{Brandt, Sascha and Jähn, Claudius and Fischer, Matthias and Meyer auf der Heide, Friedhelm}}, issn = {{0167-7055}}, journal = {{Computer Graphics Forum}}, location = {{Seoul, South Korea}}, number = {{7}}, pages = {{413--424}}, title = {{{Visibility‐Aware Progressive Farthest Point Sampling on the GPU}}}, doi = {{10.1111/cgf.13848}}, volume = {{38}}, year = {{2019}}, } @unpublished{16341, abstract = {{We present a technique for rendering highly complex 3D scenes in real-time by generating uniformly distributed points on the scene's visible surfaces. The technique is applicable to a wide range of scene types, like scenes directly based on complex and detailed CAD data consisting of billions of polygons (in contrast to scenes handcrafted solely for visualization). This allows to visualize such scenes smoothly even in VR on a HMD with good image quality, while maintaining the necessary frame-rates. In contrast to other point based rendering methods, we place points in an approximated blue noise distribution only on visible surfaces and store them in a highly GPU efficient data structure, allowing to progressively refine the number of rendered points to maximize the image quality for a given target frame rate. Our evaluation shows that scenes consisting of a high amount of polygons can be rendered with interactive frame rates with good visual quality on standard hardware.}}, author = {{Brandt, Sascha and Jähn, Claudius and Fischer, Matthias and Meyer auf der Heide, Friedhelm}}, booktitle = {{arXiv:1904.08225}}, title = {{{Rendering of Complex Heterogenous Scenes using Progressive Blue Surfels}}}, year = {{2019}}, }