
{  
   "types" : {
      "Bookmark" : {
         "pluralLabel" : "Bookmarks"
      },
      "Publication" : {
         "pluralLabel" : "Publications"
      },
      "GoldStandardPublication" : {
         "pluralLabel" : "GoldStandardPublications"
      },
      "GoldStandardBookmark" : {
         "pluralLabel" : "GoldStandardBookmarks"
      },
      "Tag" : {
         "pluralLabel" : "Tags"
      },
      "User" : {
         "pluralLabel" : "Users"
      },
      "Group" : {
         "pluralLabel" : "Groups"
      },
      "Sphere" : {
         "pluralLabel" : "Spheres"
      }
   },
   
   "properties" : {
      "count" : {
         "valueType" : "number"
      },
      "date" : {
         "valueType" : "date"
      },
      "changeDate" : {
         "valueType" : "date"
      },
      "url" : {
         "valueType" : "url"
      },
      "id" : {
         "valueType" : "url"
      },
      "tags" : {
         "valueType" : "item"
      },
      "user" : {
         "valueType" : "item"
      }      
   },
   
   "items" : [
   	  
      {
         "type" : "Publication",
         "id"   : "https://puma.ub.uni-stuttgart.de/bibtex/2681b454983a40cab8f85adcd4407c07e/simtech",         
         "tags" : [
            "EXC2075","PN7","PN7-1(II)Sedlmair","curated"
         ],
         
         "intraHash" : "681b454983a40cab8f85adcd4407c07e",
         "interHash" : "26b8366361ceedf8ea68152a0c869a35",
         "label" : "Demonstration of VisRing: A Display-Extended Smartring for Nano Visualizations",
         "user" : "simtech",
         "description" : "",
         "date" : "2026-01-16 14:13:18",
         "changeDate" : "2026-03-31 10:53:11",
         "count" : 4,
         "pub-type": "inproceedings",
         "booktitle": "Adjunct Proceedings of the 38th Annual ACM Symposium on User Interface Software and Technology","series": "UIST Adjunct '25","publisher":"Association for Computing Machinery","address":"New York, NY, USA",
         "year": "2025", 
         "url": "", 
         
         "author": [ 
            "Runze Liu","Christian Krauter","Taiting Lu","Mara Schulte","Alexander Achberger","Tanja Blascheck","Michael Sedlmair","Mahanth Gowda"
         ],
         "authors": [
         	
            	{"first" : "Runze",	"last" : "Liu"},
            	{"first" : "Christian",	"last" : "Krauter"},
            	{"first" : "Taiting",	"last" : "Lu"},
            	{"first" : "Mara",	"last" : "Schulte"},
            	{"first" : "Alexander",	"last" : "Achberger"},
            	{"first" : "Tanja",	"last" : "Blascheck"},
            	{"first" : "Michael",	"last" : "Sedlmair"},
            	{"first" : "Mahanth",	"last" : "Gowda"}
         ],
         "number": "32","pages": "4",
         "affiliation" : "Krauter, Christian, University of Stuttgart. Achberger, Alexander, Visualisierungsinstitut der Universität Stuttgart. Blascheck, Tanja, Institut für Visualisierung und Interaktive Systeme. Sedlmair, Michael, Visualisierungsinstitut der Universität Stuttgart",
         
         "orcid-numbers" : "Liu, Runze/0000-0003-2342-1644, Krauter, Christian/0000-0002-9787-0816, Lu, Taiting/0000-0002-9695-3142, Schulte, Mara/0009-0000-3575-5575, Blascheck, Tanja/0000-0003-4002-4499, Sedlmair, Michael/0000-0001-7048-9292, Gowda, Mahanth/0000-0001-5325-5013",
         
         "doi" : "10.1145/3746058.3758997",
         
         "bibtexKey": "liu:2025:demonstration"

      }
,
      {
         "type" : "Publication",
         "id"   : "https://puma.ub.uni-stuttgart.de/bibtex/2b0aea427f9003d5044304179ba59820a/simtech",         
         "tags" : [
            "EXC2075","PN7","PN7-1(II)Sedlmair","curated"
         ],
         
         "intraHash" : "b0aea427f9003d5044304179ba59820a",
         "interHash" : "5c8265c501f13abc2f0017e4eb7a37a0",
         "label" : "VisRing: A Display-Extended Smartring for Nano Visualizations",
         "user" : "simtech",
         "description" : "",
         "date" : "2026-01-16 14:12:23",
         "changeDate" : "2026-03-31 10:48:15",
         "count" : 4,
         "pub-type": "inproceedings",
         "booktitle": "Proceedings of the 38th Annual ACM Symposium on User Interface Software and Technology","series": "UIST '25","publisher":"Association for Computing Machinery","address":"New York, NY, USA",
         "year": "2025", 
         "url": "", 
         
         "author": [ 
            "Taiting Lu","Christian Krauter","Runze Liu","Mara Schulte","Alexander Achberger","Tanja Blascheck","Michael Sedlmair","Mahanth Gowda"
         ],
         "authors": [
         	
            	{"first" : "Taiting",	"last" : "Lu"},
            	{"first" : "Christian",	"last" : "Krauter"},
            	{"first" : "Runze",	"last" : "Liu"},
            	{"first" : "Mara",	"last" : "Schulte"},
            	{"first" : "Alexander",	"last" : "Achberger"},
            	{"first" : "Tanja",	"last" : "Blascheck"},
            	{"first" : "Michael",	"last" : "Sedlmair"},
            	{"first" : "Mahanth",	"last" : "Gowda"}
         ],
         "number": "111","pages": "18",
         "affiliation" : "Krauter, Christian, University of Stuttgart. Achberger, Alexander, Visualisierungsinstitut der Universität Stuttgart. Blascheck, Tanja, Institut für Visualisierung und Interaktive Systeme. Sedlmair, Michael, Visualisierungsinstitut der Universität Stuttgart",
         
         "orcid-numbers" : "Lu, Taiting/0000-0002-9695-3142, Krauter, Christian/0000-0002-9787-0816, Liu, Runze/0000-0003-2342-1644, Schulte, Mara/0009-0000-3575-5575, Blascheck, Tanja/0000-0003-4002-4499, Sedlmair, Michael/0000-0001-7048-9292, Gowda, Mahanth/0000-0001-5325-5013",
         
         "doi" : "10.1145/3746059.3747806",
         
         "bibtexKey": "lu:2025:visring:"

      }
,
      {
         "type" : "Publication",
         "id"   : "https://puma.ub.uni-stuttgart.de/bibtex/216e8d4040c5524d590d2f8b8f20cf266/simtech",         
         "tags" : [
            "EXC2075","PN7","PN7-6(II)","curated"
         ],
         
         "intraHash" : "16e8d4040c5524d590d2f8b8f20cf266",
         "interHash" : "0b53f7d1f2163f46a61fd5a6233da83b",
         "label" : "Multi-Fidelity Surrogate Model for Representing Hierarchical and Conflicting Databases to Approximate Human-Seat Interaction",
         "user" : "simtech",
         "description" : "",
         "date" : "2025-05-16 15:25:27",
         "changeDate" : "2025-10-16 15:35:53",
         "count" : 5,
         "pub-type": "article",
         "journal": "IFAC-PapersOnLine","publisher":"Elsevier BV",
         "year": "2025", 
         "url": "http://dx.doi.org/10.1016/j.ifacol.2025.03.058", 
         
         "author": [ 
            "Gia Huy Mike Huynh","Niklas Fahse","Jonas Kneifl","Joachim Linn","Jörg Fehr"
         ],
         "authors": [
         	
            	{"first" : "Gia Huy",	"last" : "Mike Huynh"},
            	{"first" : "Niklas",	"last" : "Fahse"},
            	{"first" : "Jonas",	"last" : "Kneifl"},
            	{"first" : "Joachim",	"last" : "Linn"},
            	{"first" : "Jörg",	"last" : "Fehr"}
         ],
         "volume": "59","number": "1","pages": "337\u2013342",
         "issn" : "2405-8963",
         
         "doi" : "10.1016/j.ifacol.2025.03.058",
         
         "bibtexKey": "Mike_Huynh_2025"

      }
,
      {
         "type" : "Publication",
         "id"   : "https://puma.ub.uni-stuttgart.de/bibtex/288a5d21da4a41f15c79cfacfc16236b3/simtech",         
         "tags" : [
            "EXC2075","PN7","PN7-5","curated"
         ],
         
         "intraHash" : "88a5d21da4a41f15c79cfacfc16236b3",
         "interHash" : "539e7856c2e2d296175310c13454ce03",
         "label" : "Impact of Privacy Protection Methods of Lifelogs on Remembered Memories",
         "user" : "simtech",
         "description" : "",
         "date" : "2025-02-17 14:54:08",
         "changeDate" : "2025-02-28 16:08:06",
         "count" : 7,
         "pub-type": "inproceedings",
         "booktitle": "Proc. ACM SIGCHI Conference on Human Factors in Computing Systems (CHI)",
         "year": "2023", 
         "url": "", 
         
         "author": [ 
            "Passant Elagroudy","Mohamed Khamis","Florian Mathis","Diana Irmscher","Ekta Sood","Andreas Bulling","Albrecht Schmidt"
         ],
         "authors": [
         	
            	{"first" : "Passant",	"last" : "Elagroudy"},
            	{"first" : "Mohamed",	"last" : "Khamis"},
            	{"first" : "Florian",	"last" : "Mathis"},
            	{"first" : "Diana",	"last" : "Irmscher"},
            	{"first" : "Ekta",	"last" : "Sood"},
            	{"first" : "Andreas",	"last" : "Bulling"},
            	{"first" : "Albrecht",	"last" : "Schmidt"}
         ],
         "pages": "1--10","note": "spotlight","abstract": "Lifelogging is traditionally used for memory augmentation. However, recent research shows that users\u2019 trust in the completeness and accuracy of lifelogs might skew their memories. Privacy-protection alterations such as body blurring and content deletion are commonly applied to photos to circumvent capturing sensitive information. However, their impact on how users remember memories remain unclear. To this end, we conduct a white-hat memory attack and report on an iterative experiment (N=21) to compare the impact of viewing 1) unaltered lifelogs, 2) blurred lifelogs, and 3) a subset of the lifelogs after deleting private ones, on confidently remembering memories. Findings indicate that all the privacy methods impact memories\u2019 quality similarly and that users tend to change their answers in recognition more than recall scenarios. Results also show that users have high confidence in their remembered content across all privacy methods. Our work raises awareness about the mindful designing of technological interventions.",
         "doi" : "10.1145/3544548.3581565",
         
         "bibtexKey": "elagroudy23_chi"

      }
,
      {
         "type" : "Publication",
         "id"   : "https://puma.ub.uni-stuttgart.de/bibtex/2f38edfcdce37f45530abaf7145d5a8d9/simtech",         
         "tags" : [
            "EXC2075","PN7","PN7-5","curated"
         ],
         
         "intraHash" : "f38edfcdce37f45530abaf7145d5a8d9",
         "interHash" : "091b4950a1314d694e8a6e2c5d18b823",
         "label" : "InteRead: An Eye Tracking Dataset of Interrupted Reading",
         "user" : "simtech",
         "description" : "",
         "date" : "2025-02-17 14:54:08",
         "changeDate" : "2025-02-28 16:08:06",
         "count" : 7,
         "pub-type": "inproceedings",
         "booktitle": "Proc. 31st Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING)",
         "year": "2024", 
         "url": "https://aclanthology.org/2024.lrec-main.802/", 
         
         "author": [ 
            "Francesca Zermiani","Prajit Dhar","Ekta Sood","Fabian Kögel","Andreas Bulling","Maria Wirzberger"
         ],
         "authors": [
         	
            	{"first" : "Francesca",	"last" : "Zermiani"},
            	{"first" : "Prajit",	"last" : "Dhar"},
            	{"first" : "Ekta",	"last" : "Sood"},
            	{"first" : "Fabian",	"last" : "Kögel"},
            	{"first" : "Andreas",	"last" : "Bulling"},
            	{"first" : "Maria",	"last" : "Wirzberger"}
         ],
         "pages": "9154--9169","abstract": "Eye movements during reading offer a window into cognitive processes and language comprehension, but the scarcity of reading data with interruptions -- which learners frequently encounter in their everyday learning environments -- hampers advances in the development of intelligent learning technologies. We introduce InteRead -- a novel 50-participant dataset of gaze data recorded during self-paced reading of real-world text. InteRead further offers fine-grained annotations of interruptions interspersed throughout the text as well as resumption lags incurred by these interruptions. Interruptions were triggered automatically once readers reached predefined target words. We validate our dataset by reporting interdisciplinary analyses on different measures of gaze behavior. In line with prior research, our analyses show that the interruptions as well as word length and word frequency effects significantly impact eye movements during reading. We also explore individual differences within our dataset, shedding light on the potential for tailored educational solutions. InteRead is accessible from our datasets web-page: https://www.ife.uni-stuttgart.de/en/llis/research/datasets/.",
         "bibtexKey": "zermiani24_coling"

      }
,
      {
         "type" : "Publication",
         "id"   : "https://puma.ub.uni-stuttgart.de/bibtex/2732572932bd773a1d162712578fb9ecd/simtech",         
         "tags" : [
            "EXC2075","PN7","PN7-5","curated"
         ],
         
         "intraHash" : "732572932bd773a1d162712578fb9ecd",
         "interHash" : "4f5017faa6f0f522d719d5ce15205a93",
         "label" : "Facial Composite Generation with Iterative Human Feedback",
         "user" : "simtech",
         "description" : "",
         "date" : "2025-02-17 14:54:08",
         "changeDate" : "2025-02-28 16:08:06",
         "count" : 6,
         "pub-type": "inproceedings",
         "booktitle": "Proc. The 1st Gaze Meets ML workshop, PMLR","series": "Proceedings of Machine Learning Research","publisher":"PMLR",
         "year": "2023", 
         "url": "https://proceedings.mlr.press/v210/strohm23a.html", 
         
         "author": [ 
            "Florian Strohm","Ekta Sood","Dominike Thomas","Mihai Bâce","Andreas Bulling"
         ],
         "authors": [
         	
            	{"first" : "Florian",	"last" : "Strohm"},
            	{"first" : "Ekta",	"last" : "Sood"},
            	{"first" : "Dominike",	"last" : "Thomas"},
            	{"first" : "Mihai",	"last" : "Bâce"},
            	{"first" : "Andreas",	"last" : "Bulling"}
         ],
         
         "editor": [ 
            "Ismini Lourentzou","Joy Wu","Satyananda Kashyap","Alexandros Karargyris","Leo Anthony Celi","Ban Kawas","Sachin Talathi"
         ],
         "editors": [
         	
            	{"first" : "Ismini",	"last" : "Lourentzou"},
            	{"first" : "Joy",	"last" : "Wu"},
            	{"first" : "Satyananda",	"last" : "Kashyap"},
            	{"first" : "Alexandros",	"last" : "Karargyris"},
            	{"first" : "Leo Anthony",	"last" : "Celi"},
            	{"first" : "Ban",	"last" : "Kawas"},
            	{"first" : "Sachin",	"last" : "Talathi"}
         ],
         "volume": "210","pages": "165--183","abstract": "We propose the first method in which human and AI collaborate to iteratively reconstruct the human\u2019s mental image of another person\u2019s face only from their eye gaze. Current tools for generating digital human faces involve a tedious and time-consuming manual design process. While gaze-based mental image reconstruction represents a promising alternative, previous methods still assumed prior knowledge about the target face, thereby severely limiting their practical usefulness. The key novelty of our method is a collaborative, it- erative query engine: Based on the user\u2019s gaze behaviour in each iteration, our method predicts which images to show to the user in the next iteration. Results from two human studies (N=12 and N=22) show that our method can visually reconstruct digital faces that are more similar to the mental image, and is more usable compared to other methods. As such, our findings point at the significant potential of human-AI collaboration for recon- structing mental images, potentially also beyond faces, and of human gaze as a rich source of information and a powerful mediator in said collaboration.",
         "pdf" : "https://proceedings.mlr.press/v210/strohm23a/strohm23a.pdf",
         
         "bibtexKey": "strohm23_gmml"

      }
,
      {
         "type" : "Publication",
         "id"   : "https://puma.ub.uni-stuttgart.de/bibtex/22332687c0dcf57f8a4e6bfc4adde675c/simtech",         
         "tags" : [
            "PN7","PN7-5","curated","exc2075"
         ],
         
         "intraHash" : "2332687c0dcf57f8a4e6bfc4adde675c",
         "interHash" : "1d3abae0619209fa4857a347bb8db61c",
         "label" : "VQA-MHUG: A gaze dataset to study multimodal neural attention in VQA",
         "user" : "simtech",
         "description" : "",
         "date" : "2025-02-17 14:54:08",
         "changeDate" : "2025-02-28 16:08:06",
         "count" : 6,
         "pub-type": "inproceedings",
         "booktitle": "Proc. ACL SIGNLL Conference on Computational Natural Language Learning (CoNLL)","publisher":"Association for Computational Linguistics",
         "year": "2021", 
         "url": "", 
         
         "author": [ 
            "Ekta Sood","Fabian Kögel","Florian Strohm","Prajit Dhar","Andreas Bulling"
         ],
         "authors": [
         	
            	{"first" : "Ekta",	"last" : "Sood"},
            	{"first" : "Fabian",	"last" : "Kögel"},
            	{"first" : "Florian",	"last" : "Strohm"},
            	{"first" : "Prajit",	"last" : "Dhar"},
            	{"first" : "Andreas",	"last" : "Bulling"}
         ],
         "pages": "27--43","note": "spotlight","abstract": "We present VQA-MHUG - a novel 49-participant dataset of multimodal human gaze on both images and questions during visual question answering (VQA) collected using a high-speed eye tracker. We use our dataset to analyze the similarity between human and neural attentive strategies learned by five state-of-the-art VQA models: Modulated Co-Attention Network (MCAN) with either grid or region features, Pythia, Bilinear Attention Network (BAN), and the Multimodal Factorized Bilinear Pooling Network (MFB). While prior work has focused on studying the image modality, our analyses show - for the first time - that for all models, higher correlation with human attention on text is a significant predictor of VQA performance. This finding points at a potential for improving VQA performance and, at the same time, calls for further research on neural text attention mechanisms and their integration into architectures for vision and language tasks, including but potentially also beyond VQA.",
         "code" : "https://git.hcics.simtech.uni-stuttgart.de/public-projects/vqa-mhug-interpretability",
         
         "award" : "Oral presentation",
         
         "dataset" : "https://perceptualui.org/research/datasets/VQA-MHUG/",
         
         "doi" : "10.18653/v1/2021.conll-1.3",
         
         "bibtexKey": "sood21_conll"

      }
,
      {
         "type" : "Publication",
         "id"   : "https://puma.ub.uni-stuttgart.de/bibtex/289a5c7ed25d6ec5d7638bcfd2a68f910/simtech",         
         "tags" : [
            "curated","exc2075","pn7","pn7-5"
         ],
         
         "intraHash" : "89a5c7ed25d6ec5d7638bcfd2a68f910",
         "interHash" : "7e28d355a52f4f5a098453fa16f906d8",
         "label" : "Gaze-enhanced Crossmodal Embeddings for Emotion Recognition",
         "user" : "simtech",
         "description" : "",
         "date" : "2025-02-17 14:54:08",
         "changeDate" : "2025-02-28 16:08:06",
         "count" : 6,
         "pub-type": "inproceedings",
         "booktitle": "Proc. International Symposium on Eye Tracking Research and Applications (ETRA)",
         "year": "2022", 
         "url": "", 
         
         "author": [ 
            "Ahmed Abdou","Ekta Sood","Philipp Müller","Andreas Bulling"
         ],
         "authors": [
         	
            	{"first" : "Ahmed",	"last" : "Abdou"},
            	{"first" : "Ekta",	"last" : "Sood"},
            	{"first" : "Philipp",	"last" : "Müller"},
            	{"first" : "Andreas",	"last" : "Bulling"}
         ],
         "volume": "6","pages": "1--18","abstract": "Emotional expressions are inherently multimodal -- integrating facial behavior, speech, and gaze -- but their automatic recognition is often limited to a single modality, e.g. speech during a phone call. While previous work proposed crossmodal emotion embeddings to improve monomodal recognition performance, despite its importance, a representation of gaze was not included. We propose a new approach to emotion recognition that incorporates an explicit representation of gaze in a crossmodal emotion embedding framework. We show that our method outperforms the previous state of the art for both audio-only and video-only emotion classification on the popular One-Minute Gradual Emotion Recognition dataset. Furthermore, we report extensive ablation experiments and provide insights into the performance of different state-of-the-art gaze representations and integration strategies. Our results not only underline the importance of gaze for emotion recognition but also demonstrate a practical and highly effective approach to leveraging gaze information for this task.",
         "code" : "https://git.hcics.simtech.uni-stuttgart.de/public-projects/gaze-enhanced-crossmodal-embeddings-for-emotion-recognition",
         
         "doi" : "10.1145/3530879",
         
         "bibtexKey": "abdou22_etra"

      }
,
      {
         "type" : "Publication",
         "id"   : "https://puma.ub.uni-stuttgart.de/bibtex/2c75873ce735b3cc04365c6cc575964c3/simtech",         
         "tags" : [
            "PN7","PN7-5","curated","exc2075"
         ],
         
         "intraHash" : "c75873ce735b3cc04365c6cc575964c3",
         "interHash" : "9c4b0990c6ef60dd915e4c6dd33d8a7b",
         "label" : "Video Language Co-Attention with Multimodal Fast-Learning Feature Fusion for VideoQA",
         "user" : "simtech",
         "description" : "",
         "date" : "2025-02-17 14:54:08",
         "changeDate" : "2025-02-28 16:08:06",
         "count" : 7,
         "pub-type": "inproceedings",
         "booktitle": "Proceedings of the 7th Workshop on Representation Learning for NLP","publisher":"Association for Computational Linguistics","address":"Stroudsburg",
         "year": "2022", 
         "url": "", 
         
         "author": [ 
            "Adnen Abdessaied","Ekta Sood","Andreas Bulling"
         ],
         "authors": [
         	
            	{"first" : "Adnen",	"last" : "Abdessaied"},
            	{"first" : "Ekta",	"last" : "Sood"},
            	{"first" : "Andreas",	"last" : "Bulling"}
         ],
         "pages": "143-155",
         "venue" : "Dublin, Ireland and Online",
         
         "isbn" : "978-1-955917-48-3",
         
         "research-areas" : "Computer Science",
         
         "language" : "eng",
         
         "eventdate" : "2022-05-26",
         
         "eventtitle" : "7th Workshop on Representation Learning for NLP (RepL4NLP 2022)",
         
         "affiliation" : "Abdessaied, A (Corresponding Author), Univ Stuttgart, Inst Visualizat & Interact Syst VIS, Stuttgart, Germany.\n   Abdessaied, Adnen; Sood, Ekta; Bulling, Andreas, Univ Stuttgart, Inst Visualizat & Interact Syst VIS, Stuttgart, Germany.",
         
         "unique-id" : "WOS:000847242200013",
         
         "doi" : "10.18653/v1/2022.repl4nlp-1.15",
         
         "bibtexKey": "abdessaied2022video"

      }
,
      {
         "type" : "Publication",
         "id"   : "https://puma.ub.uni-stuttgart.de/bibtex/2d5ddc756c2b5f9a2cf500b26bec8aa2a/simtech",         
         "tags" : [
            "EXC2075","PN7","PN7-5","curated"
         ],
         
         "intraHash" : "d5ddc756c2b5f9a2cf500b26bec8aa2a",
         "interHash" : "af6bd1a5a36a93d625a1f8c83b975539",
         "label" : "Multimodal Integration of Human-Like Attention in Visual Question Answering",
         "user" : "simtech",
         "description" : "",
         "date" : "2025-02-17 14:54:07",
         "changeDate" : "2025-02-28 16:08:06",
         "count" : 6,
         "pub-type": "techreport",
         
         "year": "2021", 
         "url": "https://arxiv.org/pdf/2109.13139.pdf", 
         
         "author": [ 
            "Ekta Sood","Fabian Kögel","Philipp Müller","Dominike Thomas","Mihai Bâce","Andreas Bulling"
         ],
         "authors": [
         	
            	{"first" : "Ekta",	"last" : "Sood"},
            	{"first" : "Fabian",	"last" : "Kögel"},
            	{"first" : "Philipp",	"last" : "Müller"},
            	{"first" : "Dominike",	"last" : "Thomas"},
            	{"first" : "Mihai",	"last" : "Bâce"},
            	{"first" : "Andreas",	"last" : "Bulling"}
         ],
         "pages": "1--11","note": "arxiv:2109.13139","abstract": "Human-like attention as a supervisory signal to guide neural attention has shown significant promise but is currently limited to uni-modal integration \u2013 even for inherently multi-modal tasks such as visual question answering (VQA). We present the Multimodal Human-like Attention Network (MULAN) \u2013 the first method for multimodal integration of human-like attention on image and text during training of VQA models. MULAN integrates attention predictions from two state-of-the-art text and image saliency models into neural self-attention layers of a recent transformer-based VQA model. Through evaluations on the challenging VQAv2 dataset, we show that MULAN achieves a new state-of-the-art performance of 73.98% accuracy on test-std and 73.72% on test-dev and, at the same time, has approximately 80% fewer trainable parameters than prior work. Overall, our work underlines the potential of integrating multimodal human-like and neural attention for VQA.",
         "bibtexKey": "sood21_arxiv"

      }
,
      {
         "type" : "Publication",
         "id"   : "https://puma.ub.uni-stuttgart.de/bibtex/25f0b7060bff3df6fe3393b50d82945b2/simtech",         
         "tags" : [
            "EXC2075","PN7","PN7-5","curated"
         ],
         
         "intraHash" : "5f0b7060bff3df6fe3393b50d82945b2",
         "interHash" : "38ecc86042c56ce4c08a5405bd36ee67",
         "label" : "Neural Photofit : Gaze-based Mental Image Reconstruction",
         "user" : "simtech",
         "description" : "",
         "date" : "2025-02-17 14:54:07",
         "changeDate" : "2025-02-28 16:08:06",
         "count" : 8,
         "pub-type": "inproceedings",
         "booktitle": "2021 IEEE/CVF International Conference on Computer Vision (ICCV)","publisher":"IEEE","address":"Piscataway",
         "year": "2021", 
         "url": "", 
         
         "author": [ 
            "Florian Strohm","Ekta Sood","Sven Mayer","Philipp Müller","Mihai Bâce","Andreas Bulling"
         ],
         "authors": [
         	
            	{"first" : "Florian",	"last" : "Strohm"},
            	{"first" : "Ekta",	"last" : "Sood"},
            	{"first" : "Sven",	"last" : "Mayer"},
            	{"first" : "Philipp",	"last" : "Müller"},
            	{"first" : "Mihai",	"last" : "Bâce"},
            	{"first" : "Andreas",	"last" : "Bulling"}
         ],
         "pages": "245-254",
         "venue" : "Online",
         
         "isbn" : "978-1-6654-2812-5 and 978-1-6654-2813-2",
         
         "research-areas" : "Computer Science",
         
         "language" : "eng",
         
         "eventdate" : "2021-10-10/2021-10-17",
         
         "eventtitle" : "2021 IEEE/CVF International Conference on Computer Vision (ICCV)",
         
         "affiliation" : "Strohm, F (Corresponding Author), Univ Stuttgart, Stuttgart, Germany.\n   Strohm, Florian; Sood, Ekta; Bace, Mihai; Bulling, Andreas, Univ Stuttgart, Stuttgart, Germany.\n   Mayer, Sven, Ludwig Maximilians Univ Munchen, Munich, Germany.\n   Mueller, Philipp, German Res Ctr Artificial Intelligence DFKI, Saarbrucken, Germany.",
         
         "unique-id" : "WOS:000797698900025",
         
         "doi" : "10.1109/ICCV48922.2021.00031",
         
         "bibtexKey": "strohm2021neural"

      }
,
      {
         "type" : "Publication",
         "id"   : "https://puma.ub.uni-stuttgart.de/bibtex/2116525f8369c3bf70a00a497ae4363ef/simtech",         
         "tags" : [
            "EXC2075","PN7","PN7-5","curated"
         ],
         
         "intraHash" : "116525f8369c3bf70a00a497ae4363ef",
         "interHash" : "9039d069dd9d3f940693c0f44272cb3f",
         "label" : "Improving Neural Saliency Prediction with a Cognitive Model of Human Visual Attention",
         "user" : "simtech",
         "description" : "",
         "date" : "2025-02-17 14:54:07",
         "changeDate" : "2025-02-28 16:08:06",
         "count" : 18,
         "pub-type": "inproceedings",
         "booktitle": "Proc. the 45th Annual Meeting of the Cognitive Science Society (CogSci)",
         "year": "2023", 
         "url": "", 
         
         "author": [ 
            "Ekta Sood","Lei Shi","Matteo Bortoletto","Yao Wang","Philipp Müller","Andreas Bulling"
         ],
         "authors": [
         	
            	{"first" : "Ekta",	"last" : "Sood"},
            	{"first" : "Lei",	"last" : "Shi"},
            	{"first" : "Matteo",	"last" : "Bortoletto"},
            	{"first" : "Yao",	"last" : "Wang"},
            	{"first" : "Philipp",	"last" : "Müller"},
            	{"first" : "Andreas",	"last" : "Bulling"}
         ],
         "pages": "3639--3646","note": "spotlight","abstract": "We present a novel method for saliency prediction that leverages a cognitive model of visual attention as an inductive bias. This approach is in stark contrast to recent purely data-driven saliency models that achieve performance improvements mainly by increased capacity, resulting in high computational costs and the need for large-scale training datasets. We demonstrate that by using a cognitive model, our method achieves competitive performance to the state of the art across several natural image datasets while only requiring a fraction of the parameters. Furthermore, we set the new state of the art for saliency prediction on information visualizations, demonstrating the effectiveness of our approach for cross-domain generalization. We further provide augmented versions of the full MSCOCO dataset with synthetic gaze data using the cognitive model, which we used to pre-train our method. Our results are highly promising and underline the significant potential of bridging between cognitive and data-driven models, potentially also beyond attention.",
         "code" : "https://git.hcics.simtech.uni-stuttgart.de/public-projects/neural-saliency-prediction-with-a-cognitive-model/",
         
         "supp" : "Yes",
         
         "dataset" : "https://perceptualui.org/research/datasets/MSCOCOEMMAFigureQAEMMA/",
         
         "bibtexKey": "sood23_cogsci"

      }
,
      {
         "type" : "Publication",
         "id"   : "https://puma.ub.uni-stuttgart.de/bibtex/2e2f0ae48c1a0d414a11da06b8707ccbe/simtech",         
         "tags" : [
            "EXC2075","PN7","PN7-6","PN7-6(II)","curated"
         ],
         
         "intraHash" : "e2f0ae48c1a0d414a11da06b8707ccbe",
         "interHash" : "648f8a10564fc13b79c2cf851ffc8bde",
         "label" : "Low-dimensional data-based surrogate model of a continuum-mechanical musculoskeletal system based on non-intrusive model order reduction",
         "user" : "simtech",
         "description" : "",
         "date" : "2025-02-12 23:09:28",
         "changeDate" : "2026-02-16 15:06:37",
         "count" : 10,
         "pub-type": "article",
         "journal": "Archive of Applied Mechanics",
         "year": "2023", 
         "url": "https://doi.org/10.1007/s00419-023-02458-5", 
         
         "author": [ 
            "Jonas Kneifl","David Rosin","Okan Avci","Oliver Röhrle","Jörg Fehr"
         ],
         "authors": [
         	
            	{"first" : "Jonas",	"last" : "Kneifl"},
            	{"first" : "David",	"last" : "Rosin"},
            	{"first" : "Okan",	"last" : "Avci"},
            	{"first" : "Oliver",	"last" : "Röhrle"},
            	{"first" : "Jörg",	"last" : "Fehr"}
         ],
         "volume": "93","number": "9","pages": "3637--3663","abstract": "Over the last decades, computer modeling has evolved from a supporting tool for engineering prototype design to an ubiquitous instrument in non-traditional fields such as medical rehabilitation. This area comes with unique challenges, e.g. the complex modeling of soft tissue or the analysis of musculoskeletal systems. Conventional modeling approaches like the finite element (FE) method are computationally costly when dealing with such models, limiting their usability for real-time simulation or deployment on low-end hardware, if the model at hand cannot be simplified without losing its expressiveness. Non-traditional approaches such as surrogate modeling using data-driven model order reduction are used to make complex high-fidelity models more widely available regardless. They often involve a dimensionality reduction step, in which the high-dimensional system state is transformed onto a low-dimensional subspace or manifold, and a regression approach to capture the reduced system behavior. While most publications focus on one dimensionality reduction, such as principal component analysis (PCA) (linear) or autoencoder (nonlinear), we consider and compare PCA, kernel PCA, autoencoders, as well as variational autoencoders for the approximation of a continuum-mechanical system. In detail, we demonstrate the benefits of the surrogate modeling approach on a complex musculoskeletal system of a human upper-arm with severe nonlinearities and physiological geometry. We consider both, the model\u2019s deformation and the internal stress as the two main quantities of interest in a FE context. By doing so we are able to create computationally low-cost surrogate models which capture the system behavior with high approximation quality and fast evaluations.",
         "issn" : "1432-0681",
         
         "refid" : "Kneifl2023",
         
         "doi" : "10.1007/s00419-023-02458-5",
         
         "bibtexKey": "kneifl2023lowdimensional"

      }
,
      {
         "type" : "Publication",
         "id"   : "https://puma.ub.uni-stuttgart.de/bibtex/248a1b2c6536c32577fdb9a0da66e0286/simtech",         
         "tags" : [
            "EXC2075","PN3","PN3-2(II)","PN7","PN7-6(II)","curated"
         ],
         
         "intraHash" : "48a1b2c6536c32577fdb9a0da66e0286",
         "interHash" : "dec5e4de3c09fe0f1a1d965be19d8310",
         "label" : "Accelerated Non\u2010linear Stability Analysis Based on Predictions From Data\u2010Based Surrogate Models",
         "user" : "simtech",
         "description" : "",
         "date" : "2025-01-03 13:57:10",
         "changeDate" : "2025-02-14 11:23:50",
         "count" : 6,
         "pub-type": "article",
         "journal": "International Journal for Numerical Methods in Engineering","publisher":"Wiley",
         "year": "2024", 
         "url": "http://dx.doi.org/10.1002/nme.7649", 
         
         "author": [ 
            "Anika Strauß","Jonas Kneifl","Anton Tkachuk","Jörg Fehr","Manfred Bischoff"
         ],
         "authors": [
         	
            	{"first" : "Anika",	"last" : "Strauß"},
            	{"first" : "Jonas",	"last" : "Kneifl"},
            	{"first" : "Anton",	"last" : "Tkachuk"},
            	{"first" : "Jörg",	"last" : "Fehr"},
            	{"first" : "Manfred",	"last" : "Bischoff"}
         ],
         "volume": "126","number": "1",
         "issn" : "1097-0207",
         
         "doi" : "10.1002/nme.7649",
         
         "bibtexKey": "Strau__2024"

      }
,
      {
         "type" : "Publication",
         "id"   : "https://puma.ub.uni-stuttgart.de/bibtex/28f7a86877dff9d2b910f088e0fd1944e/simtech",         
         "tags" : [
            "EXC2075","PN7","PN7-1(II)Sedlmair","PN7-1.3","curated"
         ],
         
         "intraHash" : "8f7a86877dff9d2b910f088e0fd1944e",
         "interHash" : "6b3aab26f50d7372a065566893fee27b",
         "label" : "PerSiVal: On-Body AR Visualization of Biomechanical Arm Simulations",
         "user" : "simtech",
         "description" : "",
         "date" : "2024-12-12 10:33:58",
         "changeDate" : "2025-02-25 11:35:30",
         "count" : 6,
         "pub-type": "article",
         "journal": "IEEE Computer Graphics and Applications",
         "year": "2024", 
         "url": "https://doi.org/10.1109/MCG.2024.3494598", 
         
         "author": [ 
            "Xingyao Yu","David Rosin","Johannes Kässinger","Benjamin Lee","Frank Dürr","Christian Becker","Oliver Röhrle","Michael Sedlmair"
         ],
         "authors": [
         	
            	{"first" : "Xingyao",	"last" : "Yu"},
            	{"first" : "David",	"last" : "Rosin"},
            	{"first" : "Johannes",	"last" : "Kässinger"},
            	{"first" : "Benjamin",	"last" : "Lee"},
            	{"first" : "Frank",	"last" : "Dürr"},
            	{"first" : "Christian",	"last" : "Becker"},
            	{"first" : "Oliver",	"last" : "Röhrle"},
            	{"first" : "Michael",	"last" : "Sedlmair"}
         ],
         "pages": "1-14","abstract": "In this work, we explore different combinations of techniques for an interactive, on-body visualization in augmented reality (AR) of an upper arm muscle simulation model. In terms of data, we focus on a continuum-mechanical simulation model involving five different muscles of the human upper arm, with physiologically realistic geometry. In terms of use cases, we focus on the immersive illustration, education, and dissemination of such simulation models. We describe the process of developing six on-body visualization prototypes over the period of five years. For each prototype, we employed different types of motion capture, AR display technologies, and visual encoding approaches, and gathered feedback throughout outreach activities. We reflect on the development of the individual prototypes and summarize lessons learned of our exploration process into the design space of situated on-body visualization.",
         "issn" : "1558-1756",
         
         "language" : "English",
         
         "doi" : "10.1109/MCG.2024.3494598",
         
         "bibtexKey": "10747503"

      }
,
      {
         "type" : "Publication",
         "id"   : "https://puma.ub.uni-stuttgart.de/bibtex/206617cdd3b123aca9801eb5205b09f1f/simtech",         
         "tags" : [
            "curated","PN7-6(II)","PN7","PN7-6","EXC2075"
         ],
         
         "intraHash" : "06617cdd3b123aca9801eb5205b09f1f",
         "interHash" : "2ffaae3d8e2209369441c9740b61edf8",
         "label" : "Multi-hierarchical surrogate learning for explicit structural dynamical systems using graph convolutional neural networks",
         "user" : "simtech",
         "description" : "",
         "date" : "2024-10-15 16:58:51",
         "changeDate" : "2025-02-12 22:54:26",
         "count" : 7,
         "pub-type": "article",
         "journal": "Computational Mechanics",
         "year": "2024", 
         "url": "https://doi.org/10.1007/s00466-024-02553-6", 
         
         "author": [ 
            "Jonas Kneifl","Jörg Fehr","Steven L. Brunton","J. Nathan Kutz"
         ],
         "authors": [
         	
            	{"first" : "Jonas",	"last" : "Kneifl"},
            	{"first" : "Jörg",	"last" : "Fehr"},
            	{"first" : "Steven L.",	"last" : "Brunton"},
            	{"first" : "J. Nathan",	"last" : "Kutz"}
         ],
         "abstract": "Highly nonlinear dynamic finite element simulations using explicit time integration are particularly valuable tools for structural analysis in fields like automotive, aerospace, and civil engineering, or in the study of injury biomechanics. However, such state-of-the-art simulation models demand significant computational resources. Conventional data-driven surrogate modeling approaches address this by evolving the dynamics on low-dimensional embeddings, yet the majority of them operate directly on high-resolution data obtained from numerical discretizations, making them costly and unsuitable for adaptive resolutions or for handling information flow over large spatial distances. We therefore propose a multi-hierarchical framework for the structured creation of a series of surrogate models at different resolutions. Macroscale features are captured on coarse surrogates, while microscale effects are resolved on finer ones, while leveraging transfer learning to pass information between scales. The objective of this study is to develop efficient surrogates for a kart frame model in a frontal impact scenario. To achieve this, its mesh is simplified to obtain multi-resolution representations of the kart. Subsequently, a graph-convolutional neural network-based surrogate learns parameter-dependent low-dimensional latent dynamics on the coarsest representation. Following surrogates are trained on residuals using finer resolutions, allowing for multiple surrogates with varying hardware requirements and increasing accuracy.",
         "issn" : "1432-0924",
         
         "preprinturl" : "https://doi.org/10.48550/arXiv.2402.09234",
         
         "doi" : "10.1007/s00466-024-02553-6",
         
         "bibtexKey": "Kneifl2024"

      }
,
      {
         "type" : "Publication",
         "id"   : "https://puma.ub.uni-stuttgart.de/bibtex/2c2d0b618b4e46ed049c92ab1a4f09ae2/simtech",         
         "tags" : [
            "PNX","PN7","EXC2075"
         ],
         
         "intraHash" : "c2d0b618b4e46ed049c92ab1a4f09ae2",
         "interHash" : "24686d34e22b161ee931faf9c77f1a63",
         "label" : "Promoting sustainable behavior: addressing user clusters through targeted incentives",
         "user" : "simtech",
         "description" : "",
         "date" : "2024-09-25 09:40:54",
         "changeDate" : "2024-10-04 08:49:24",
         "count" : 4,
         "pub-type": "article",
         "journal": "Humanities and Social Sciences Communications",
         "year": "2024", 
         "url": "http://dx.doi.org/10.1057/s41599-024-03581-6", 
         
         "author": [ 
            "Laura Höpfl","Maximilian Grimlitza","Isabella Lang","Maria Wirzberger"
         ],
         "authors": [
         	
            	{"first" : "Laura",	"last" : "Höpfl"},
            	{"first" : "Maximilian",	"last" : "Grimlitza"},
            	{"first" : "Isabella",	"last" : "Lang"},
            	{"first" : "Maria",	"last" : "Wirzberger"}
         ],
         "abstract": "<jats:title>Abstract</jats:title><jats:p>Given the urgency of climate change action and the significant climate impact of household emissions, understanding the drivers of individuals\u2019 sustainable behavior patterns is more important than ever. Consequently, we investigate whether different clusters of individual users can be distinguished regarding sustainability-related values, attitudes, and intentions. If these diverse clusters exist, we can explore tailored approaches to promote sustainable behavior patterns among them based on their unique needs and targets. For this purpose, we employ a mixed-method approach combining qualitative interviews with a quantitative survey. The obtained insights help us identify core factors that drive sustainable behavior, develop representations of different user groups, and suggest individualized interventions for supporting sustainable behavior patterns. The qualitative part comprised interviews with ten participants, resulting in the development of qualitative personas. Emerging differences could subsequently be used to select validated psychological scales for the quantitative part to confirm the differences. Applying data-driven clustering, we identify five intention-based clusters that vary regarding factors such as belief in climate change, collaboration, or skepticism concerning sustainability. Building on both qualitative and quantitative insights, five validated personas are created for research and practical use. These personas include Socially Sustainable, Responsible Savers, Unconcerned Spenders, Comfort-Oriented, and Skeptical Consumers. Individuals corresponding to the selected persona may, for example, respond positively to sustainability benefits, while others may be more receptive to hedonistic benefits. Addressing related varying motivational factors raises the demand for individualized interventions. These could be achieved by incorporating the personas\u2019 needs with more individualized products and services to promote sustainable behavior.</jats:p>",
         "issn" : "2662-9992",
         
         "doi" : "10.1057/s41599-024-03581-6",
         
         "bibtexKey": "hopfl2024promoting"

      }
,
      {
         "type" : "Publication",
         "id"   : "https://puma.ub.uni-stuttgart.de/bibtex/2b1084a07104d9d1132be8af4581360ab/simtech",         
         "tags" : [
            "EXC2075","PN7","PN7-1(II)Sedlmair","PN7-1.3","curated"
         ],
         
         "intraHash" : "b1084a07104d9d1132be8af4581360ab",
         "interHash" : "6c2799500d702cde816a3ccc26ef2dc5",
         "label" : "ChoreoVis: Planning and Assessing Formations in Dance Choreographies",
         "user" : "simtech",
         "description" : "",
         "date" : "2024-09-17 14:30:20",
         "changeDate" : "2024-10-07 09:24:50",
         "count" : 7,
         "pub-type": "proceedings",
         "journal": "Computer Graphics Forum","publisher":"The Eurographics Association and John Wiley & Sons Ltd.",
         "year": "2024", 
         "url": "https://diglib.eg.org/handle/10.1111/cgf15104", 
         
         "author": [ 
            "Samuel Beck","Nina Doerr","Kuno Kurzhals","Alexander Riedlinger","Fabian Schmierer","Michael Sedlmair","Steffen Koch"
         ],
         "authors": [
         	
            	{"first" : "Samuel",	"last" : "Beck"},
            	{"first" : "Nina",	"last" : "Doerr"},
            	{"first" : "Kuno",	"last" : "Kurzhals"},
            	{"first" : "Alexander",	"last" : "Riedlinger"},
            	{"first" : "Fabian",	"last" : "Schmierer"},
            	{"first" : "Michael",	"last" : "Sedlmair"},
            	{"first" : "Steffen",	"last" : "Koch"}
         ],
         "volume": "43","number": "3","abstract": "Sports visualization has developed into an active research field over the last decades. Many approaches focus on analyzing movement data recorded from unstructured situations, such as soccer. For the analysis of choreographed activities like formation dancing, however, the goal differs, as dancers follow specific formations in coordinated movement trajectories. To date, little work exists on how visual analytics methods can support such choreographed performances. To fill this gap, we introduce a new visual approach for planning and assessing dance choreographies. In terms of planning choreographies, we contribute a web application with interactive authoring tools and views for the dancers' positions and orientations, movement trajectories, poses, dance floor utilization, and movement distances. For assessing dancers' real-world movement trajectories, extracted by manual bounding box annotations, we developed a timeline showing aggregated trajectory deviations and a dance floor view for detailed trajectory comparison. Our approach was developed and evaluated in collaboration with dance instructors, showing that introducing visual analytics into this domain promises improvements in training efficiency for the future.",
         "issn" : "1467-8659",
         
         "doi" : "10.1111/cgf.15104",
         
         "bibtexKey": "10.1111:cgf.15104"

      }
,
      {
         "type" : "Publication",
         "id"   : "https://puma.ub.uni-stuttgart.de/bibtex/25d2f808133b0b5ef1694fa776b46177d/simtech",         
         "tags" : [
            "EXC2075","PN7","PN7-1(II)Sedlmair","PN7-1.3","curated"
         ],
         
         "intraHash" : "5d2f808133b0b5ef1694fa776b46177d",
         "interHash" : "0f4c31756eb9bd0ca1340bcb7b99756a",
         "label" : "Sitting Posture Recognition and Feedback: A Literature Review",
         "user" : "simtech",
         "description" : "",
         "date" : "2024-09-17 14:29:19",
         "changeDate" : "2024-10-07 09:24:50",
         "count" : 8,
         "pub-type": "inproceedings",
         "booktitle": "CHI '24: Proceedings of the CHI Conference on Human Factors in Computing Systems","series": "CHI","publisher":"ACM","address":"NY",
         "year": "2024", 
         "url": "https://github.com/ChristianKrauter/sitting-posture-review-material", 
         
         "author": [ 
            "Christian Krauter","Katrin Angerbauer","Aimée Sousa Calepso","Alexander Achberger","Sven Mayer","Michael Sedlmair"
         ],
         "authors": [
         	
            	{"first" : "Christian",	"last" : "Krauter"},
            	{"first" : "Katrin",	"last" : "Angerbauer"},
            	{"first" : "Aimée",	"last" : "Sousa Calepso"},
            	{"first" : "Alexander",	"last" : "Achberger"},
            	{"first" : "Sven",	"last" : "Mayer"},
            	{"first" : "Michael",	"last" : "Sedlmair"}
         ],
         
         "affiliation" : "Krauter, Christian, University of Stuttgart. Angerbauer, Katrin, Visualisierungsinstitut der Universität Stuttgart. Sousa Calepso, Aimée, Visualisierungsinstitut der Universität Stuttgart. Achberger, Alexander, Visualisierungsinstitut der Universität Stuttgart. Sedlmair, Michael, Visualisierungsinstitut der Universität Stuttgart",
         
         "orcid-numbers" : "Krauter, Christian/0000-0002-9787-0816, Angerbauer, Katrin/0000-0002-1126-5288, Sedlmair, Michael/0000-0001-7048-9292",
         
         "doi" : "10.1145/3613904.3642657",
         
         "bibtexKey": "krauter:2024:sitting"

      }
,
      {
         "type" : "Publication",
         "id"   : "https://puma.ub.uni-stuttgart.de/bibtex/256098df22d41a8c64c592971e155a9d5/simtech",         
         "tags" : [
            "EXC2075","PN7","PN7-1(II)Becker","PN7-1.1","curated"
         ],
         
         "intraHash" : "56098df22d41a8c64c592971e155a9d5",
         "interHash" : "c457add8aa0d32d7b146f5c6f0c7661c",
         "label" : "Persival: Simulating Complex 3D Meshes on Resource-Constrained Mobile AR Devices Using Interpolation",
         "user" : "simtech",
         "description" : "",
         "date" : "2024-09-08 23:02:23",
         "changeDate" : "2024-09-16 14:38:49",
         "count" : 11,
         "pub-type": "inproceedings",
         "booktitle": "Proceedings of the 2022 IEEE International Conference on Distributed Computing Systems (ICDCS)",
         "year": "2022", 
         "url": "", 
         
         "author": [ 
            "Johannes Kässinger","David Rosin","Frank Dürr","Niklas Hornischer","Kurt Rothermel","Oliver Röhrle"
         ],
         "authors": [
         	
            	{"first" : "Johannes",	"last" : "Kässinger"},
            	{"first" : "David",	"last" : "Rosin"},
            	{"first" : "Frank",	"last" : "Dürr"},
            	{"first" : "Niklas",	"last" : "Hornischer"},
            	{"first" : "Kurt",	"last" : "Rothermel"},
            	{"first" : "Oliver",	"last" : "Röhrle"}
         ],
         "pages": "961--971",
         "doi" : "10.1109/ICDCS54860.2022.00097",
         
         "bibtexKey": "kassinger2022persival"

      }
	  
   ]
}
