David M. Schmidt

@dmschmidt
40 Followers
109 Following
20 Posts
#NLProc PhD Student & Research Associate at Bielefeld University
Working on: Question Answering over Linked Data, Semantic Web, Lexical Knowledge & Compositionality in AI
Websitehttps://davidmschmidt.de
I am incredibly happy to share that our paper "CompoST: A Benchmark for Analyzing the Ability of LLMs To Compositionally Interpret Questions in a QALD Setting" has been accepted as a research track paper at the International Semantic Web Conference @iswc_conf! Huge thanks to my co-authors Raoul Schubert and Philipp Cimiano! Stay tuned for the paper and see you all in Japan!
A huge thanks to everyone who made this week such a unique, memorable experience! And, if you are Master's/PhD student or PostDoc, I cannot recommend too much to apply for the next iteration of #isws !
- worked in a research task force on building a reliable LLM-based metadata enrichment pipeline for cultural heritage objects (special thanks to our tutor Valentina Presutti and our whole team), as well as writing a corresponding white paper and presenting our results in the final session
- got to know and discussed with so many fascinating people
- and listened to keynotes and tutorials of well-known researchers such as Frank van Harmelen, Natasha Noy and Enrico Motta
The #ISWS2025 experience really managed to combine lots of fun activities and challenges (that we took an oath to keep secret 🤫), working with leading figures of the Semantic Web field as well as intense networking in a unique, wonderful way! It felt like a month worth of program items and activities had been compressed to one magnificent piece of art.
During the last week, among many other things, I
- summarized the motivation of my work in a 45s "Minute Madness" session
What a week! I just had the incredible opportunity to attend the International Semantic Web Research Summer School 2025 @isws in Bertinoro, Italy. I hoped for an intense week filled with inspiring keynotes, interesting people to talk with and opportunities to present my work on Question Answering over Linked Data and compositionality - and I got so much more than "just" that!
The objective of this study was to evaluate to what extent state-of-the-art large language models can appropriately summarize posts shared by patients in web-based forums and health communities. Specifically, the goal was to compare the performance of different language models and prompting strategies on the task of summarizing documents reflecting the experiences of individual patients.
Social media is acknowledged by regulatory bodies (e.g., the Food and Drug Administration) as an important source of patient experience data to learn about patients’ unmet needs, priorities, and preferences. However, current methods rely either on manual analysis and do not scale, or on automatic processing, yielding mainly quantitative insights. Methods that can automatically summarize texts and yield qualitative insights at scale are missing.

🚀 New paper! 🚀

I am happy to announce our paper "Summarizing Online Patient Conversations Using Generative Language Models: Experimental and Comparative Study," which has just been published in JMIR Medical Informatics!

🎓 Authors: Rakhi Asokkumar Subjagouri Nair, Matthias Hartung, Philipp Heinisch, Janik Jaskolski, Cornelius Starke-Knäusel, Susana Veríssimo, David M. Schmidt, Philipp Cimiano

🔗 Paper: https://doi.org/10.2196/62909

Summarizing Online Patient Conversations Using Generative Language Models: Experimental and Comparative Study

Background: Social media is acknowledged by regulatory bodies (eg, the Food and Drug Administration) as an important source of patient experience data to learn about patients’ unmet needs, priorities, and preferences. However, current methods rely either on manual analysis and do not scale, or on automatic processing, yielding mainly quantitative insights. Methods that can automatically summarize texts and yield qualitative insights at scale are missing. Objective: The objective of this study was to evaluate to what extent state-of-the-art large language models can appropriately summarize posts shared by patients in web-based forums and health communities. Specifically, the goal was to compare the performance of different language models and prompting strategies on the task of summarizing documents reflecting the experiences of individual patients. Methods: In our experimental and comparative study, we applied 3 different language models (Flan-T5, Generative Pretrained Transformer [GPT], GPT-3, and GPT-3.5) in combination with various prompting strategies to the task of summarizing posts from patients in online communities. The generated summaries were evaluated with respect to 124 manually created summaries as a ground-truth reference. As evaluation metrics, we used 2 standard metrics from the field of text generation, namely, Recall-Oriented Understudy for Gisting Evaluation (ROUGE) and BERTScore, to compare the automatically generated summaries to the manually created reference summaries. Results: Among the zero-shot prompting–based large language models investigated, GPT-3.5 performed better than the other models with respect to the ROUGE metrics, as well as with respect to BERTScore. While zero-shot prompting seems to be a good prompting strategy, overall GPT-3.5 in combination with directional stimulus prompting in a 3-shot setting had the best results with respect to the aforementioned metrics. A manual investigation of the summarization of the best-performing method showed that the generated summaries were accurate and plausible compared to the manual summaries. Conclusions: Taken together, our results suggest that state-of-the-art pretrained language models are a valuable tool to provide qualitative insights about the patient experience to better understand unmet needs, patient priorities, and how a disease impacts daily functioning and quality of life to inform processes aimed at improving health care delivery and ensure that drug development focuses more on the actual priorities and unmet needs of patients. The key limitations of our work are the small data sample as well as the fact that the manual summaries were created by 1 annotator only. Furthermore, the results hold only for the examined models and prompting strategies, potentially not generalizing to other models and strategies.

JMIR Medical Informatics

We currently have two fully-funded open PhD positions in our group with a focus on #NLProc, #InformationExtraction and #TextGeneration. I can really recommend both the group as well as Philipp Cimiano as a supervisor, so take this opportunity!

NLP/Text Generation
EN: https://uni-bielefeld.hr4you.org/job/view/4054
DE: https://uni-bielefeld.hr4you.org/job/view/4053

NLP/Information Extraction
EN: https://uni-bielefeld.hr4you.org/job/view/4059
DE: https://uni-bielefeld.hr4you.org/job/view/4057

If you have any questions, do not hesitate to contact me or Philipp directly!

Research Position - Text Generation/Natural Langua...

<div style="text-align: justify;">The Faculty of Engineering at Bielefeld University is looking for a research assistant to work on th...

🚀 We are #hiring! Are you interested in Natural Language Processing, Text Generation or Information Extraction and want to pursue a PhD?

Then you now have the chance to become a part of the Semantic Computing Group at Bielefeld University!

Application Deadline: 20.03.2025