<?xml version="1.0" encoding="UTF-8"?>
<rdf:RDF xmlns:community="http://www.bibsonomy.org/ontologies/2008/05/community#" xmlns:foaf="http://xmlns.com/foaf/0.1/" xmlns:owl="http://www.w3.org/2002/07/owl#" xmlns:admin="http://webns.net/mvcb/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:syn="http://purl.org/rss/1.0/modules/syndication/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" xmlns:cc="http://web.resource.org/cc/" xmlns:xsd="http://www.w3.org/2001/XMLSchema#" xmlns:swrc="http://swrc.ontoware.org/ontology#" xmlns:rdfs="http://www.w3.org/2000/01/rdf-schema#" xmlns="http://purl.org/rss/1.0/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xml:base="https://puma.ub.uni-stuttgart.de/group/simtech_test"><owl:Ontology rdf:about=""><rdfs:comment>PUMA publications for /group/simtech_test</rdfs:comment><owl:imports rdf:resource="http://swrc.ontoware.org/ontology/portal"/></owl:Ontology><rdf:Description rdf:about="https://puma.ub.uni-stuttgart.de/bibtex/2a3c99a8a008a1c87735c6a17b70d04dd/janrange"><owl:sameAs rdf:resource="/uri/bibtex/2a3c99a8a008a1c87735c6a17b70d04dd/janrange"/><rdf:type rdf:resource="http://swrc.ontoware.org/ontology#Article"/><owl:sameAs rdf:resource="http://dx.doi.org/10.1016/j.envsoft.2025.106792"/><swrc:date>Fri Feb 27 18:01:12 CET 2026</swrc:date><swrc:journal>Environmental Modelling &amp; Software</swrc:journal><swrc:month>04</swrc:month><swrc:pages>106792</swrc:pages><swrc:publisher><swrc:Organization swrc:name="Elsevier BV"/></swrc:publisher><swrc:title>Advancing geospatial data infrastructure in Dataverse via metadata automation, interactive tools and LLM case study</swrc:title><swrc:volume>199</swrc:volume><swrc:year>2026</swrc:year><swrc:keywords>myown rdm fair </swrc:keywords><swrc:abstract>In the era of big data and interdisciplinary research, the effective dissemination and reuse of geospatial data have become vital across various fields such as economics, biostatistics, epidemiology, environmental health, and sciences. This study investigates the challenges associated with managing geospatial data and presents the implementation of tools designed to address these challenges. We present an overview of the current state of geospatial data in a general-purpose research data repository Dataverse and outline a series of implemented advancements for improving the management and utilization of geospatial datasets. These advancements include building the capability to extract structured metadata automatically, enabling programmatic engagement with data assets, incorporating checklists, facilitating geospatial-specific searches, and providing previews of geographic dataset coverage. In this paper, we include two case studies. In the first, we evaluate the effectiveness of the automatic metadata extraction feature, part of our proposed advancements, using the large language model GPT-4 and find that the extracted metadata offers unique information, which is not typically provided by the user. In the second case study, we introduce the community of practice around climate-health data at Dataverse, coordinated through the CAFE Research Coordinating Center.</swrc:abstract><swrc:hasExtraField><swrc:Field swrc:value="English" swrc:key="language"/></swrc:hasExtraField><swrc:hasExtraField><swrc:Field swrc:value="1364-8152" swrc:key="issn"/></swrc:hasExtraField><swrc:hasExtraField><swrc:Field swrc:value="10.1016/j.envsoft.2025.106792" swrc:key="doi"/></swrc:hasExtraField><swrc:author><rdf:Seq><rdf:_1><swrc:Person swrc:name="Ana Trišović"/></rdf:_1><rdf:_2><swrc:Person swrc:name="Jan Range"/></rdf:_2><rdf:_3><swrc:Person swrc:name="Philip Durbin"/></rdf:_3><rdf:_4><swrc:Person swrc:name="Katherine Mika"/></rdf:_4><rdf:_5><swrc:Person swrc:name="Amber Leahey"/></rdf:_5><rdf:_6><swrc:Person swrc:name="Wei Li"/></rdf:_6></rdf:Seq></swrc:author></rdf:Description><rdf:Description rdf:about="https://puma.ub.uni-stuttgart.de/bibtex/2d5ddc756c2b5f9a2cf500b26bec8aa2a/hermann"><owl:sameAs rdf:resource="/uri/bibtex/2d5ddc756c2b5f9a2cf500b26bec8aa2a/hermann"/><rdf:type rdf:resource="http://swrc.ontoware.org/ontology#TechnicalReport"/><owl:sameAs rdf:resource="https://arxiv.org/pdf/2109.13139.pdf"/><swrc:date>Mon Feb 17 14:54:07 CET 2025</swrc:date><swrc:month>10</swrc:month><swrc:note>arxiv:2109.13139</swrc:note><swrc:pages>1--11</swrc:pages><swrc:title>Multimodal Integration of Human-Like Attention in Visual Question Answering</swrc:title><swrc:year>2021</swrc:year><swrc:keywords>pn7 updated pn7-5 exc2075 </swrc:keywords><swrc:abstract>Human-like attention as a supervisory signal to guide neural attention has shown significant promise but is currently limited to uni-modal integration – even for inherently multi-modal tasks such as visual question answering (VQA). We present the Multimodal Human-like Attention Network (MULAN) – the first method for multimodal integration of human-like attention on image and text during training of VQA models. MULAN integrates attention predictions from two state-of-the-art text and image saliency models into neural self-attention layers of a recent transformer-based VQA model. Through evaluations on the challenging VQAv2 dataset, we show that MULAN achieves a new state-of-the-art performance of 73.98% accuracy on test-std and 73.72% on test-dev and, at the same time, has approximately 80% fewer trainable parameters than prior work. Overall, our work underlines the potential of integrating multimodal human-like and neural attention for VQA.</swrc:abstract><swrc:author><rdf:Seq><rdf:_1><swrc:Person swrc:name="Ekta Sood"/></rdf:_1><rdf:_2><swrc:Person swrc:name="Fabian Kögel"/></rdf:_2><rdf:_3><swrc:Person swrc:name="Philipp Müller"/></rdf:_3><rdf:_4><swrc:Person swrc:name="Dominike Thomas"/></rdf:_4><rdf:_5><swrc:Person swrc:name="Mihai Bâce"/></rdf:_5><rdf:_6><swrc:Person swrc:name="Andreas Bulling"/></rdf:_6></rdf:Seq></swrc:author></rdf:Description><rdf:Description rdf:about="https://puma.ub.uni-stuttgart.de/bibtex/2732572932bd773a1d162712578fb9ecd/hermann"><owl:sameAs rdf:resource="/uri/bibtex/2732572932bd773a1d162712578fb9ecd/hermann"/><rdf:type rdf:resource="http://swrc.ontoware.org/ontology#InProceedings"/><owl:sameAs rdf:resource="https://proceedings.mlr.press/v210/strohm23a.html"/><swrc:date>Mon Feb 17 14:54:07 CET 2025</swrc:date><swrc:booktitle>Proc. The 1st Gaze Meets ML workshop, PMLR</swrc:booktitle><swrc:month>03 Dec</swrc:month><swrc:pages>165--183</swrc:pages><swrc:publisher><swrc:Organization swrc:name="PMLR"/></swrc:publisher><swrc:series>Proceedings of Machine Learning Research</swrc:series><swrc:title>Facial Composite Generation with Iterative Human Feedback</swrc:title><swrc:volume>210</swrc:volume><swrc:year>2023</swrc:year><swrc:keywords>pn7 updated pn7-5 exc2075 </swrc:keywords><swrc:abstract>We propose the first method in which human and AI collaborate to iteratively reconstruct the human’s mental image of another person’s face only from their eye gaze. Current tools for generating digital human faces involve a tedious and time-consuming manual design process. While gaze-based mental image reconstruction represents a promising alternative, previous methods still assumed prior knowledge about the target face, thereby severely limiting their practical usefulness. The key novelty of our method is a collaborative, it- erative query engine: Based on the user’s gaze behaviour in each iteration, our method predicts which images to show to the user in the next iteration. Results from two human studies (N=12 and N=22) show that our method can visually reconstruct digital faces that are more similar to the mental image, and is more usable compared to other methods. As such, our findings point at the significant potential of human-AI collaboration for recon- structing mental images, potentially also beyond faces, and of human gaze as a rich source of information and a powerful mediator in said collaboration.</swrc:abstract><swrc:hasExtraField><swrc:Field swrc:value="https://proceedings.mlr.press/v210/strohm23a/strohm23a.pdf" swrc:key="pdf"/></swrc:hasExtraField><swrc:author><rdf:Seq><rdf:_1><swrc:Person swrc:name="Florian Strohm"/></rdf:_1><rdf:_2><swrc:Person swrc:name="Ekta Sood"/></rdf:_2><rdf:_3><swrc:Person swrc:name="Dominike Thomas"/></rdf:_3><rdf:_4><swrc:Person swrc:name="Mihai B{\^a}ce"/></rdf:_4><rdf:_5><swrc:Person swrc:name="Andreas Bulling"/></rdf:_5></rdf:Seq></swrc:author><swrc:editor><rdf:Seq><rdf:_1><swrc:Person swrc:name="Ismini Lourentzou"/></rdf:_1><rdf:_2><swrc:Person swrc:name="Joy Wu"/></rdf:_2><rdf:_3><swrc:Person swrc:name="Satyananda Kashyap"/></rdf:_3><rdf:_4><swrc:Person swrc:name="Alexandros Karargyris"/></rdf:_4><rdf:_5><swrc:Person swrc:name="Leo Anthony Celi"/></rdf:_5><rdf:_6><swrc:Person swrc:name="Ban Kawas"/></rdf:_6><rdf:_7><swrc:Person swrc:name="Sachin Talathi"/></rdf:_7></rdf:Seq></swrc:editor></rdf:Description><rdf:Description rdf:about="https://puma.ub.uni-stuttgart.de/bibtex/2a6b4a6f7774f1abad2001d0bfaf39f02/hermann"><owl:sameAs rdf:resource="/uri/bibtex/2a6b4a6f7774f1abad2001d0bfaf39f02/hermann"/><rdf:type rdf:resource="http://swrc.ontoware.org/ontology#InProceedings"/><owl:sameAs rdf:resource="https://aclanthology.org/2020.conll-1.2"/><swrc:date>Mon Feb 17 14:54:07 CET 2025</swrc:date><swrc:address>Online</swrc:address><swrc:booktitle>Proceedings of the 24th Conference on Computational Natural Language Learning</swrc:booktitle><swrc:month>11</swrc:month><swrc:pages>12--25</swrc:pages><swrc:publisher><swrc:Organization swrc:name="Association for Computational Linguistics"/></swrc:publisher><swrc:title>Interpreting Attention Models with Human Visual Attention in Machine Reading Comprehension</swrc:title><swrc:year>2020</swrc:year><swrc:keywords>pn7 updated pn7-5 exc2075 </swrc:keywords><swrc:abstract>While neural networks with attention mechanisms have achieved superior performance on many natural language processing tasks, it remains unclear to which extent learned attention resembles human visual attention. In this paper, we propose a new method that leverages eye-tracking data to investigate the relationship between human visual attention and neural attention in machine reading comprehension. To this end, we introduce a novel 23 participant eye tracking dataset - MQA-RC, in which participants read movie plots and answered pre-defined questions. We compare state of the art networks based on long short-term memory (LSTM), convolutional neural models (CNN) and XLNet Transformer architectures. We find that higher similarity to human attention and performance significantly correlates to the LSTM and CNN models. However, we show this relationship does not hold true for the XLNet models {--} despite the fact that the XLNet performs best on this challenging task. Our results suggest that different architectures seem to learn rather different neural attention strategies and similarity of neural to human attention does not guarantee best performance.</swrc:abstract><swrc:hasExtraField><swrc:Field swrc:value="10.18653/v1/2020.conll-1.2" swrc:key="doi"/></swrc:hasExtraField><swrc:author><rdf:Seq><rdf:_1><swrc:Person swrc:name="Ekta Sood"/></rdf:_1><rdf:_2><swrc:Person swrc:name="Simon Tannert"/></rdf:_2><rdf:_3><swrc:Person swrc:name="Diego Frassinelli"/></rdf:_3><rdf:_4><swrc:Person swrc:name="Andreas Bulling"/></rdf:_4><rdf:_5><swrc:Person swrc:name="Ngoc Thang Vu"/></rdf:_5></rdf:Seq></swrc:author></rdf:Description><rdf:Description rdf:about="https://puma.ub.uni-stuttgart.de/bibtex/2f97b9935cacc514b3e94fb2318b6f61e/hermann"><owl:sameAs rdf:resource="/uri/bibtex/2f97b9935cacc514b3e94fb2318b6f61e/hermann"/><rdf:type rdf:resource="http://swrc.ontoware.org/ontology#InProceedings"/><owl:sameAs rdf:resource="https://doi.org/10.1145/3379155.3391332"/><swrc:date>Mon Feb 17 14:54:07 CET 2025</swrc:date><swrc:address>New York, NY, USA</swrc:address><swrc:booktitle>ACM Symposium on Eye Tracking Research and Applications</swrc:booktitle><swrc:month>06</swrc:month><swrc:pages>1–10</swrc:pages><swrc:publisher><swrc:Organization swrc:name="Association for Computing Machinery"/></swrc:publisher><swrc:series>ETRA &#039;20 Full Papers</swrc:series><swrc:title>Anticipating Averted Gaze in Dyadic Interactions</swrc:title><swrc:year>2020</swrc:year><swrc:keywords>pn7 updated pn7-5 exc2075 </swrc:keywords><swrc:day>2</swrc:day><swrc:abstract>We present the first method to anticipate averted gaze in natural dyadic interactions.
The task of anticipating averted gaze, i.e. that a person will not make eye contact
in the near future, remains unsolved despite its importance for human social encounters
as well as a number of applications, including human-robot interaction or conversational
agents. Our multimodal method is based on a long short-term memory (LSTM) network
that analyses non-verbal facial cues and speaking behaviour. We empirically evaluate
our method for different future time horizons on a novel dataset of 121 YouTube videos
of dyadic video conferences (74 hours in total). We investigate person-specific and
person-independent performance and demonstrate that our method clearly outperforms
baselines in both settings. As such, our work sheds light on the tight interplay between
eye contact and other non-verbal signals and underlines the potential of computational
modelling and anticipation of averted gaze for interactive applications.</swrc:abstract><swrc:hasExtraField><swrc:Field swrc:value="9781450371339" swrc:key="isbn"/></swrc:hasExtraField><swrc:hasExtraField><swrc:Field swrc:value="Stuttgart, Germany" swrc:key="location"/></swrc:hasExtraField><swrc:hasExtraField><swrc:Field swrc:value="10.1145/3379155.3391332" swrc:key="doi"/></swrc:hasExtraField><swrc:author><rdf:Seq><rdf:_1><swrc:Person swrc:name="Philipp Müller"/></rdf:_1><rdf:_2><swrc:Person swrc:name="Ekta Sood"/></rdf:_2><rdf:_3><swrc:Person swrc:name="Andreas Bulling"/></rdf:_3></rdf:Seq></swrc:author></rdf:Description><rdf:Description rdf:about="https://puma.ub.uni-stuttgart.de/bibtex/2116525f8369c3bf70a00a497ae4363ef/hermann"><owl:sameAs rdf:resource="/uri/bibtex/2116525f8369c3bf70a00a497ae4363ef/hermann"/><rdf:type rdf:resource="http://swrc.ontoware.org/ontology#InProceedings"/><swrc:date>Mon Feb 17 14:54:07 CET 2025</swrc:date><swrc:booktitle>Proc. the 45th Annual Meeting of the Cognitive Science Society (CogSci)</swrc:booktitle><swrc:month>07</swrc:month><swrc:note>spotlight</swrc:note><swrc:pages>3639--3646</swrc:pages><swrc:title>Improving Neural Saliency Prediction with a Cognitive Model of Human Visual Attention</swrc:title><swrc:year>2023</swrc:year><swrc:keywords>pn7 updated pn7-5 exc2075 </swrc:keywords><swrc:abstract>We present a novel method for saliency prediction that leverages a cognitive model of visual attention as an inductive bias. This approach is in stark contrast to recent purely data-driven saliency models that achieve performance improvements mainly by increased capacity, resulting in high computational costs and the need for large-scale training datasets. We demonstrate that by using a cognitive model, our method achieves competitive performance to the state of the art across several natural image datasets while only requiring a fraction of the parameters. Furthermore, we set the new state of the art for saliency prediction on information visualizations, demonstrating the effectiveness of our approach for cross-domain generalization. We further provide augmented versions of the full MSCOCO dataset with synthetic gaze data using the cognitive model, which we used to pre-train our method. Our results are highly promising and underline the significant potential of bridging between cognitive and data-driven models, potentially also beyond attention.</swrc:abstract><swrc:hasExtraField><swrc:Field swrc:value="https://git.hcics.simtech.uni-stuttgart.de/public-projects/neural-saliency-prediction-with-a-cognitive-model/" swrc:key="code"/></swrc:hasExtraField><swrc:hasExtraField><swrc:Field swrc:value="Yes" swrc:key="supp"/></swrc:hasExtraField><swrc:hasExtraField><swrc:Field swrc:value="https://perceptualui.org/research/datasets/MSCOCOEMMAFigureQAEMMA/" swrc:key="dataset"/></swrc:hasExtraField><swrc:author><rdf:Seq><rdf:_1><swrc:Person swrc:name="Ekta Sood"/></rdf:_1><rdf:_2><swrc:Person swrc:name="Lei Shi"/></rdf:_2><rdf:_3><swrc:Person swrc:name="Matteo Bortoletto"/></rdf:_3><rdf:_4><swrc:Person swrc:name="Yao Wang"/></rdf:_4><rdf:_5><swrc:Person swrc:name="Philipp Müller"/></rdf:_5><rdf:_6><swrc:Person swrc:name="Andreas Bulling"/></rdf:_6></rdf:Seq></swrc:author></rdf:Description><rdf:Description rdf:about="https://puma.ub.uni-stuttgart.de/bibtex/25f0b7060bff3df6fe3393b50d82945b2/hermann"><owl:sameAs rdf:resource="/uri/bibtex/25f0b7060bff3df6fe3393b50d82945b2/hermann"/><rdf:type rdf:resource="http://swrc.ontoware.org/ontology#InProceedings"/><swrc:date>Mon Feb 17 14:54:07 CET 2025</swrc:date><swrc:address>Piscataway</swrc:address><swrc:booktitle>2021 IEEE/CVF International Conference on Computer Vision (ICCV)</swrc:booktitle><swrc:pages>245-254</swrc:pages><swrc:publisher><swrc:Organization swrc:name="IEEE"/></swrc:publisher><swrc:title>Neural Photofit : Gaze-based Mental Image Reconstruction</swrc:title><swrc:year>2021</swrc:year><swrc:keywords>pn7 updated pn7-5 exc2075 </swrc:keywords><swrc:hasExtraField><swrc:Field swrc:value="Online" swrc:key="venue"/></swrc:hasExtraField><swrc:hasExtraField><swrc:Field swrc:value="{978-1-6654-2812-5} and {978-1-6654-2813-2}" swrc:key="isbn"/></swrc:hasExtraField><swrc:hasExtraField><swrc:Field swrc:value="Computer Science" swrc:key="research-areas"/></swrc:hasExtraField><swrc:hasExtraField><swrc:Field swrc:value="eng" swrc:key="language"/></swrc:hasExtraField><swrc:hasExtraField><swrc:Field swrc:value="2021-10-10/2021-10-17" swrc:key="eventdate"/></swrc:hasExtraField><swrc:hasExtraField><swrc:Field swrc:value="2021 IEEE/CVF International Conference on Computer Vision (ICCV)" swrc:key="eventtitle"/></swrc:hasExtraField><swrc:hasExtraField><swrc:Field swrc:value="Strohm, F (Corresponding Author), Univ Stuttgart, Stuttgart, Germany.
   Strohm, Florian; Sood, Ekta; Bace, Mihai; Bulling, Andreas, Univ Stuttgart, Stuttgart, Germany.
   Mayer, Sven, Ludwig Maximilians Univ Munchen, Munich, Germany.
   Mueller, Philipp, German Res Ctr Artificial Intelligence DFKI, Saarbrucken, Germany." swrc:key="affiliation"/></swrc:hasExtraField><swrc:hasExtraField><swrc:Field swrc:value="WOS:000797698900025" swrc:key="unique-id"/></swrc:hasExtraField><swrc:hasExtraField><swrc:Field swrc:value="10.1109/ICCV48922.2021.00031" swrc:key="doi"/></swrc:hasExtraField><swrc:author><rdf:Seq><rdf:_1><swrc:Person swrc:name="Florian Strohm"/></rdf:_1><rdf:_2><swrc:Person swrc:name="Ekta Sood"/></rdf:_2><rdf:_3><swrc:Person swrc:name="Sven Mayer"/></rdf:_3><rdf:_4><swrc:Person swrc:name="Philipp Müller"/></rdf:_4><rdf:_5><swrc:Person swrc:name="Mihai Bâce"/></rdf:_5><rdf:_6><swrc:Person swrc:name="Andreas Bulling"/></rdf:_6></rdf:Seq></swrc:author></rdf:Description><rdf:Description rdf:about="https://puma.ub.uni-stuttgart.de/bibtex/288a5d21da4a41f15c79cfacfc16236b3/hermann"><owl:sameAs rdf:resource="/uri/bibtex/288a5d21da4a41f15c79cfacfc16236b3/hermann"/><rdf:type rdf:resource="http://swrc.ontoware.org/ontology#InProceedings"/><swrc:date>Mon Feb 17 14:54:07 CET 2025</swrc:date><swrc:booktitle>Proc. ACM SIGCHI Conference on Human Factors in Computing Systems (CHI)</swrc:booktitle><swrc:note>spotlight</swrc:note><swrc:pages>1--10</swrc:pages><swrc:title>Impact of Privacy Protection Methods of Lifelogs on Remembered Memories</swrc:title><swrc:year>2023</swrc:year><swrc:keywords>pn7 updated pn7-5 exc2075 </swrc:keywords><swrc:abstract>Lifelogging is traditionally used for memory augmentation. However, recent research shows that users’ trust in the completeness and accuracy of lifelogs might skew their memories. Privacy-protection alterations such as body blurring and content deletion are commonly applied to photos to circumvent capturing sensitive information. However, their impact on how users remember memories remain unclear. To this end, we conduct a white-hat memory attack and report on an iterative experiment (N=21) to compare the impact of viewing 1) unaltered lifelogs, 2) blurred lifelogs, and 3) a subset of the lifelogs after deleting private ones, on confidently remembering memories. Findings indicate that all the privacy methods impact memories’ quality similarly and that users tend to change their answers in recognition more than recall scenarios. Results also show that users have high confidence in their remembered content across all privacy methods. Our work raises awareness about the mindful designing of technological interventions.</swrc:abstract><swrc:hasExtraField><swrc:Field swrc:value="10.1145/3544548.3581565" swrc:key="doi"/></swrc:hasExtraField><swrc:author><rdf:Seq><rdf:_1><swrc:Person swrc:name="Passant Elagroudy"/></rdf:_1><rdf:_2><swrc:Person swrc:name="Mohamed Khamis"/></rdf:_2><rdf:_3><swrc:Person swrc:name="Florian Mathis"/></rdf:_3><rdf:_4><swrc:Person swrc:name="Diana Irmscher"/></rdf:_4><rdf:_5><swrc:Person swrc:name="Ekta Sood"/></rdf:_5><rdf:_6><swrc:Person swrc:name="Andreas Bulling"/></rdf:_6><rdf:_7><swrc:Person swrc:name="Albrecht Schmidt"/></rdf:_7></rdf:Seq></swrc:author></rdf:Description><rdf:Description rdf:about="https://puma.ub.uni-stuttgart.de/bibtex/289a5c7ed25d6ec5d7638bcfd2a68f910/hermann"><owl:sameAs rdf:resource="/uri/bibtex/289a5c7ed25d6ec5d7638bcfd2a68f910/hermann"/><rdf:type rdf:resource="http://swrc.ontoware.org/ontology#InProceedings"/><swrc:date>Mon Feb 17 14:54:07 CET 2025</swrc:date><swrc:booktitle>Proc. International Symposium on Eye Tracking Research and Applications (ETRA)</swrc:booktitle><swrc:pages>1--18</swrc:pages><swrc:title>Gaze-enhanced Crossmodal Embeddings for Emotion Recognition</swrc:title><swrc:volume>6</swrc:volume><swrc:year>2022</swrc:year><swrc:keywords>pn7 updated pn7-5 exc2075 </swrc:keywords><swrc:abstract>Emotional expressions are inherently multimodal -- integrating facial behavior, speech, and gaze -- but their automatic recognition is often limited to a single modality, e.g. speech during a phone call. While previous work proposed crossmodal emotion embeddings to improve monomodal recognition performance, despite its importance, a representation of gaze was not included. We propose a new approach to emotion recognition that incorporates an explicit representation of gaze in a crossmodal emotion embedding framework. We show that our method outperforms the previous state of the art for both audio-only and video-only emotion classification on the popular One-Minute Gradual Emotion Recognition dataset. Furthermore, we report extensive ablation experiments and provide insights into the performance of different state-of-the-art gaze representations and integration strategies. Our results not only underline the importance of gaze for emotion recognition but also demonstrate a practical and highly effective approach to leveraging gaze information for this task.</swrc:abstract><swrc:hasExtraField><swrc:Field swrc:value="https://git.hcics.simtech.uni-stuttgart.de/public-projects/gaze-enhanced-crossmodal-embeddings-for-emotion-recognition" swrc:key="code"/></swrc:hasExtraField><swrc:hasExtraField><swrc:Field swrc:value="10.1145/3530879" swrc:key="doi"/></swrc:hasExtraField><swrc:author><rdf:Seq><rdf:_1><swrc:Person swrc:name="Ahmed Abdou"/></rdf:_1><rdf:_2><swrc:Person swrc:name="Ekta Sood"/></rdf:_2><rdf:_3><swrc:Person swrc:name="Philipp Müller"/></rdf:_3><rdf:_4><swrc:Person swrc:name="Andreas Bulling"/></rdf:_4></rdf:Seq></swrc:author></rdf:Description><rdf:Description rdf:about="https://puma.ub.uni-stuttgart.de/bibtex/22332687c0dcf57f8a4e6bfc4adde675c/hermann"><owl:sameAs rdf:resource="/uri/bibtex/22332687c0dcf57f8a4e6bfc4adde675c/hermann"/><rdf:type rdf:resource="http://swrc.ontoware.org/ontology#InProceedings"/><swrc:date>Mon Feb 17 14:54:07 CET 2025</swrc:date><swrc:booktitle>Proc. ACL SIGNLL Conference on Computational Natural Language Learning (CoNLL)</swrc:booktitle><swrc:month>11</swrc:month><swrc:note>spotlight</swrc:note><swrc:pages>27--43</swrc:pages><swrc:publisher><swrc:Organization swrc:name="Association for Computational Linguistics"/></swrc:publisher><swrc:title>VQA-MHUG: A gaze dataset to study multimodal neural attention in VQA</swrc:title><swrc:year>2021</swrc:year><swrc:keywords>pn7 updated pn7-5 exc2075 </swrc:keywords><swrc:abstract>We present VQA-MHUG - a novel 49-participant dataset of multimodal human gaze on both images and questions during visual question answering (VQA) collected using a high-speed eye tracker. We use our dataset to analyze the similarity between human and neural attentive strategies learned by five state-of-the-art VQA models: Modulated Co-Attention Network (MCAN) with either grid or region features, Pythia, Bilinear Attention Network (BAN), and the Multimodal Factorized Bilinear Pooling Network (MFB). While prior work has focused on studying the image modality, our analyses show - for the first time - that for all models, higher correlation with human attention on text is a significant predictor of VQA performance. This finding points at a potential for improving VQA performance and, at the same time, calls for further research on neural text attention mechanisms and their integration into architectures for vision and language tasks, including but potentially also beyond VQA.</swrc:abstract><swrc:hasExtraField><swrc:Field swrc:value="https://git.hcics.simtech.uni-stuttgart.de/public-projects/vqa-mhug-interpretability" swrc:key="code"/></swrc:hasExtraField><swrc:hasExtraField><swrc:Field swrc:value="Oral presentation" swrc:key="award"/></swrc:hasExtraField><swrc:hasExtraField><swrc:Field swrc:value="https://perceptualui.org/research/datasets/VQA-MHUG/" swrc:key="dataset"/></swrc:hasExtraField><swrc:hasExtraField><swrc:Field swrc:value="10.18653/v1/2021.conll-1.3" swrc:key="doi"/></swrc:hasExtraField><swrc:author><rdf:Seq><rdf:_1><swrc:Person swrc:name="Ekta Sood"/></rdf:_1><rdf:_2><swrc:Person swrc:name="Fabian Kögel"/></rdf:_2><rdf:_3><swrc:Person swrc:name="Florian Strohm"/></rdf:_3><rdf:_4><swrc:Person swrc:name="Prajit Dhar"/></rdf:_4><rdf:_5><swrc:Person swrc:name="Andreas Bulling"/></rdf:_5></rdf:Seq></swrc:author></rdf:Description><rdf:Description rdf:about="https://puma.ub.uni-stuttgart.de/bibtex/2f38edfcdce37f45530abaf7145d5a8d9/hermann"><owl:sameAs rdf:resource="/uri/bibtex/2f38edfcdce37f45530abaf7145d5a8d9/hermann"/><rdf:type rdf:resource="http://swrc.ontoware.org/ontology#InProceedings"/><owl:sameAs rdf:resource="https://aclanthology.org/2024.lrec-main.802/"/><swrc:date>Mon Feb 17 14:54:07 CET 2025</swrc:date><swrc:booktitle>Proc. 31st Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING)</swrc:booktitle><swrc:pages>9154--9169</swrc:pages><swrc:title>InteRead: An Eye Tracking Dataset of Interrupted Reading</swrc:title><swrc:year>2024</swrc:year><swrc:keywords>pn7 updated pn7-5 exc2075 </swrc:keywords><swrc:abstract>Eye movements during reading offer a window into cognitive processes and language comprehension, but the scarcity of reading data with interruptions {--} which learners frequently encounter in their everyday learning environments {--} hampers advances in the development of intelligent learning technologies. We introduce InteRead {--} a novel 50-participant dataset of gaze data recorded during self-paced reading of real-world text. InteRead further offers fine-grained annotations of interruptions interspersed throughout the text as well as resumption lags incurred by these interruptions. Interruptions were triggered automatically once readers reached predefined target words. We validate our dataset by reporting interdisciplinary analyses on different measures of gaze behavior. In line with prior research, our analyses show that the interruptions as well as word length and word frequency effects significantly impact eye movements during reading. We also explore individual differences within our dataset, shedding light on the potential for tailored educational solutions. InteRead is accessible from our datasets web-page: https://www.ife.uni-stuttgart.de/en/llis/research/datasets/.</swrc:abstract><swrc:author><rdf:Seq><rdf:_1><swrc:Person swrc:name="Francesca Zermiani"/></rdf:_1><rdf:_2><swrc:Person swrc:name="Prajit Dhar"/></rdf:_2><rdf:_3><swrc:Person swrc:name="Ekta Sood"/></rdf:_3><rdf:_4><swrc:Person swrc:name="Fabian Kögel"/></rdf:_4><rdf:_5><swrc:Person swrc:name="Andreas Bulling"/></rdf:_5><rdf:_6><swrc:Person swrc:name="Maria Wirzberger"/></rdf:_6></rdf:Seq></swrc:author></rdf:Description><rdf:Description rdf:about="https://puma.ub.uni-stuttgart.de/bibtex/271dca1344ec8f44c6e672bafb003df36/hermann"><owl:sameAs rdf:resource="/uri/bibtex/271dca1344ec8f44c6e672bafb003df36/hermann"/><rdf:type rdf:resource="http://swrc.ontoware.org/ontology#InProceedings"/><owl:sameAs rdf:resource="https://proceedings.neurips.cc/paper/2020/file/460191c72f67e90150a093b4585e7eb4-Paper.pdf"/><swrc:date>Mon Feb 17 14:54:07 CET 2025</swrc:date><swrc:booktitle>Advances in Neural Information Processing Systems</swrc:booktitle><swrc:pages>6327--6341</swrc:pages><swrc:publisher><swrc:Organization swrc:name="Curran Associates, Inc."/></swrc:publisher><swrc:title>Improving Natural Language Processing Tasks with Human Gaze-Guided Neural Attention</swrc:title><swrc:volume>33</swrc:volume><swrc:year>2020</swrc:year><swrc:keywords>pn7 PN7-5 updated exc2075 </swrc:keywords><swrc:author><rdf:Seq><rdf:_1><swrc:Person swrc:name="Ekta Sood"/></rdf:_1><rdf:_2><swrc:Person swrc:name="Simon Tannert"/></rdf:_2><rdf:_3><swrc:Person swrc:name="Philipp Mueller"/></rdf:_3><rdf:_4><swrc:Person swrc:name="Andreas Bulling"/></rdf:_4></rdf:Seq></swrc:author><swrc:editor><rdf:Seq><rdf:_1><swrc:Person swrc:name="H. Larochelle"/></rdf:_1><rdf:_2><swrc:Person swrc:name="M. Ranzato"/></rdf:_2><rdf:_3><swrc:Person swrc:name="R. Hadsell"/></rdf:_3><rdf:_4><swrc:Person swrc:name="M. F. Balcan"/></rdf:_4><rdf:_5><swrc:Person swrc:name="H. Lin"/></rdf:_5></rdf:Seq></swrc:editor></rdf:Description><rdf:Description rdf:about="https://puma.ub.uni-stuttgart.de/bibtex/2c75873ce735b3cc04365c6cc575964c3/hermann"><owl:sameAs rdf:resource="/uri/bibtex/2c75873ce735b3cc04365c6cc575964c3/hermann"/><rdf:type rdf:resource="http://swrc.ontoware.org/ontology#InProceedings"/><swrc:date>Mon Feb 17 14:54:07 CET 2025</swrc:date><swrc:address>Stroudsburg</swrc:address><swrc:booktitle>Proceedings of the 7th Workshop on Representation Learning for NLP</swrc:booktitle><swrc:pages>143-155</swrc:pages><swrc:publisher><swrc:Organization swrc:name="Association for Computational Linguistics"/></swrc:publisher><swrc:title>Video Language Co-Attention with Multimodal Fast-Learning Feature Fusion for VideoQA</swrc:title><swrc:year>2022</swrc:year><swrc:keywords>pn7 updated pn7-5 exc2075 </swrc:keywords><swrc:hasExtraField><swrc:Field swrc:value="{Dublin, Ireland} and {Online}" swrc:key="venue"/></swrc:hasExtraField><swrc:hasExtraField><swrc:Field swrc:value="978-1-955917-48-3" swrc:key="isbn"/></swrc:hasExtraField><swrc:hasExtraField><swrc:Field swrc:value="Computer Science" swrc:key="research-areas"/></swrc:hasExtraField><swrc:hasExtraField><swrc:Field swrc:value="eng" swrc:key="language"/></swrc:hasExtraField><swrc:hasExtraField><swrc:Field swrc:value="2022-05-26" swrc:key="eventdate"/></swrc:hasExtraField><swrc:hasExtraField><swrc:Field swrc:value="7th Workshop on Representation Learning for NLP (RepL4NLP 2022)" swrc:key="eventtitle"/></swrc:hasExtraField><swrc:hasExtraField><swrc:Field swrc:value="Abdessaied, A (Corresponding Author), Univ Stuttgart, Inst Visualizat &amp; Interact Syst VIS, Stuttgart, Germany.
   Abdessaied, Adnen; Sood, Ekta; Bulling, Andreas, Univ Stuttgart, Inst Visualizat &amp; Interact Syst VIS, Stuttgart, Germany." swrc:key="affiliation"/></swrc:hasExtraField><swrc:hasExtraField><swrc:Field swrc:value="WOS:000847242200013" swrc:key="unique-id"/></swrc:hasExtraField><swrc:hasExtraField><swrc:Field swrc:value="10.18653/v1/2022.repl4nlp-1.15" swrc:key="doi"/></swrc:hasExtraField><swrc:author><rdf:Seq><rdf:_1><swrc:Person swrc:name="Adnen Abdessaied"/></rdf:_1><rdf:_2><swrc:Person swrc:name="Ekta Sood"/></rdf:_2><rdf:_3><swrc:Person swrc:name="Andreas Bulling"/></rdf:_3></rdf:Seq></swrc:author></rdf:Description><rdf:Description rdf:about="https://puma.ub.uni-stuttgart.de/bibtex/2316925253e231493e407c90abf096169/simtech_test"><owl:sameAs rdf:resource="/uri/bibtex/2316925253e231493e407c90abf096169/simtech_test"/><rdf:type rdf:resource="http://swrc.ontoware.org/ontology#Article"/><swrc:date>Fri Nov 08 09:56:54 CET 2024</swrc:date><swrc:month>01</swrc:month><swrc:title>Test Post Deletion Copy Version 3</swrc:title><swrc:volume>1</swrc:volume><swrc:year>2024</swrc:year><swrc:keywords>pn1 curated test </swrc:keywords><swrc:author><rdf:Seq><rdf:_1><swrc:Person swrc:name="John Doe"/></rdf:_1></rdf:Seq></swrc:author><swrc:editor><rdf:Seq><rdf:_1><swrc:Person swrc:name="John Doe"/></rdf:_1></rdf:Seq></swrc:editor></rdf:Description><rdf:Description rdf:about="https://puma.ub.uni-stuttgart.de/bibtex/2f3808778b44fb208a5d531b65f7b10bc/simtech_test"><owl:sameAs rdf:resource="/uri/bibtex/2f3808778b44fb208a5d531b65f7b10bc/simtech_test"/><rdf:type rdf:resource="http://swrc.ontoware.org/ontology#Article"/><swrc:date>Fri Nov 08 09:56:26 CET 2024</swrc:date><swrc:journal>my test journal</swrc:journal><swrc:title>Test Post Deletion Copy Version 2</swrc:title><swrc:volume>1</swrc:volume><swrc:year>2024</swrc:year><swrc:keywords>pn1 test </swrc:keywords><swrc:abstract>my test abstract</swrc:abstract><swrc:author><rdf:Seq><rdf:_1><swrc:Person swrc:name="John Doe"/></rdf:_1></rdf:Seq></swrc:author><swrc:editor><rdf:Seq><rdf:_1><swrc:Person swrc:name="John Doe"/></rdf:_1></rdf:Seq></swrc:editor></rdf:Description><rdf:Description rdf:about="https://puma.ub.uni-stuttgart.de/bibtex/2316925253e231493e407c90abf096169/fwang"><owl:sameAs rdf:resource="/uri/bibtex/2316925253e231493e407c90abf096169/fwang"/><rdf:type rdf:resource="http://swrc.ontoware.org/ontology#Article"/><swrc:date>Fri Nov 08 09:20:25 CET 2024</swrc:date><swrc:month>01</swrc:month><swrc:title>Test Post Deletion Copy Version 3</swrc:title><swrc:volume>1</swrc:volume><swrc:year>2024</swrc:year><swrc:keywords>pn1 curated test </swrc:keywords><swrc:author><rdf:Seq><rdf:_1><swrc:Person swrc:name="John Doe"/></rdf:_1></rdf:Seq></swrc:author><swrc:editor><rdf:Seq><rdf:_1><swrc:Person swrc:name="John Doe"/></rdf:_1></rdf:Seq></swrc:editor></rdf:Description><rdf:Description rdf:about="https://puma.ub.uni-stuttgart.de/bibtex/2c1829e1296a4eec6aaa9761063e0c925/fwang"><owl:sameAs rdf:resource="/uri/bibtex/2c1829e1296a4eec6aaa9761063e0c925/fwang"/><rdf:type rdf:resource="http://swrc.ontoware.org/ontology#Article"/><swrc:date>Fri Nov 08 09:12:50 CET 2024</swrc:date><swrc:journal>my updated test journal</swrc:journal><swrc:title>Test Post Deletion Copy Version 5</swrc:title><swrc:year>2024</swrc:year><swrc:keywords>pn1 </swrc:keywords><swrc:author><rdf:Seq><rdf:_1><swrc:Person swrc:name="John Doe"/></rdf:_1></rdf:Seq></swrc:author></rdf:Description><rdf:Description rdf:about="https://puma.ub.uni-stuttgart.de/bibtex/22400ade403666203352e41bae2aeda68/fwang"><owl:sameAs rdf:resource="/uri/bibtex/22400ade403666203352e41bae2aeda68/fwang"/><rdf:type rdf:resource="http://swrc.ontoware.org/ontology#Article"/><swrc:date>Tue Oct 29 10:10:04 CET 2024</swrc:date><swrc:journal>my updated test journal</swrc:journal><swrc:month>01</swrc:month><swrc:title>Test Post Deletion Copy Version 3</swrc:title><swrc:volume>1</swrc:volume><swrc:year>2024</swrc:year><swrc:keywords>pn1 curated test </swrc:keywords><swrc:abstract>my test abstract</swrc:abstract><swrc:author><rdf:Seq><rdf:_1><swrc:Person swrc:name="John Doe"/></rdf:_1></rdf:Seq></swrc:author><swrc:editor><rdf:Seq><rdf:_1><swrc:Person swrc:name="John Doe"/></rdf:_1></rdf:Seq></swrc:editor></rdf:Description><rdf:Description rdf:about="https://puma.ub.uni-stuttgart.de/bibtex/24f483a8e0b3cd726e47e5310aefa7ba5/fwang"><owl:sameAs rdf:resource="/uri/bibtex/24f483a8e0b3cd726e47e5310aefa7ba5/fwang"/><rdf:type rdf:resource="http://swrc.ontoware.org/ontology#Misc"/><swrc:date>Mon Oct 28 15:38:37 CET 2024</swrc:date><swrc:booktitle>unknown </swrc:booktitle><swrc:journal>my test journal</swrc:journal><swrc:month>01</swrc:month><swrc:number>2</swrc:number><swrc:pages>3-4</swrc:pages><swrc:title>Test Post Deletion Copy Version 5</swrc:title><swrc:volume>1</swrc:volume><swrc:year>2024</swrc:year><swrc:keywords>pn1 curated test </swrc:keywords><swrc:abstract>my updated test abstract</swrc:abstract><swrc:author><rdf:Seq><rdf:_1><swrc:Person swrc:name="John Doe"/></rdf:_1></rdf:Seq></swrc:author><swrc:editor><rdf:Seq><rdf:_1><swrc:Person swrc:name="John Doe"/></rdf:_1></rdf:Seq></swrc:editor></rdf:Description><rdf:Description rdf:about="https://puma.ub.uni-stuttgart.de/bibtex/2578a0288dff5b31c98feeb77ababfac2/fwang"><owl:sameAs rdf:resource="/uri/bibtex/2578a0288dff5b31c98feeb77ababfac2/fwang"/><rdf:type rdf:resource="http://swrc.ontoware.org/ontology#Article"/><swrc:date>Wed Oct 23 15:05:18 CEST 2024</swrc:date><swrc:booktitle>unknown </swrc:booktitle><swrc:journal>my test journal</swrc:journal><swrc:month>01</swrc:month><swrc:number>2</swrc:number><swrc:pages>3-4</swrc:pages><swrc:title>Test Post Deletion Copy Version 3</swrc:title><swrc:volume>1</swrc:volume><swrc:year>2024</swrc:year><swrc:keywords>pn1 curated test </swrc:keywords><swrc:abstract>my updated test abstract</swrc:abstract><swrc:author><rdf:Seq><rdf:_1><swrc:Person swrc:name="John Doe"/></rdf:_1></rdf:Seq></swrc:author><swrc:editor><rdf:Seq><rdf:_1><swrc:Person swrc:name="John Doe"/></rdf:_1></rdf:Seq></swrc:editor></rdf:Description></rdf:RDF>