@collaborativeai

15 Followers
1 Following
14 Posts
The Collaborative Artificial Intelligence (CAI) Group, headed by Prof. Dr. Andreas Bulling, is within the Department of Computer Science at the University of Stuttgart, Germany. Our group conducts fundamental research towards collaborative artificial intelligence at the intersection of multimodal machine learning, computational cognitive modelling, computer vision, and human-machine interaction.
Websitecollaborative-ai.org

Excited to share that our paper โ€œRobustSpring: Benchmarking Robustness to Image Corruptions for Optical Flow, Scene Flow and Stereoโ€ has been accepted to #ICLR2026 ๐ŸŽ‰.

We introduce RobustSpring, a new benchmark that evaluates not only accuracy but also robustness of optical flow, scene flow, and stereo models under 20 realโ€‘world image corruptions.

Congratulations to the authors!

๐ŸŒ For more news: https://www.collaborative-ai.org/publications/oei26_iclr/

Which part of graphs do people look at when solving analytical tasks?

๐Ÿ“ฐ Our work "Towards a Better Understanding of Graph Perception in Immersive Environments" was accepted to Graph Drawing #GD2025.

Congratulations to the authors!

Learn more about this work from our website: https://www.collaborative-ai.org/publications/zhang25_gd/

๐Ÿš€ Exciting News! ๐Ÿš€

HOIGaze: Gaze Estimation During Hand-Object Interactions in Extended Reality has been accepted to #SIGGRAPH 2025! ๐ŸŽ‰

HOIGaze introduces:1๏ธโƒฃ A hierarchical framework that first identifies which hand the user is visually attending to, then estimates gaze direction based on the handโ€™s posture. 2๏ธโƒฃ A gaze estimation network that combines graph neural networks and cross-modal Transformers. 3๏ธโƒฃ An eye-head coordination loss function.

๐Ÿ” Learn more: https://collaborative-ai.org/publications/hu25_siggraph/

Collaborative Artificial Intelligence

Our group conducts fundamental research towards collaborative artificial intelligence (CAI) at the intersection of multimodal machine learning, computational cognitive modelling, computer vision, and human-machine interaction.

๐ŸŽ‰ Exciting News! ๐ŸŽ‰

We're thrilled to share that our group's proposal, "๐— ๐˜‚๐—น๐˜๐—ถ๐— ๐—ฒ๐—ฑ๐—ถ๐—ฎ๐˜๐—ฒ: ๐— ๐˜‚๐—น๐˜๐—ถ๐—บ๐—ผ๐—ฑ๐—ฎ๐—น ๐—•๐—ฒ๐—ต๐—ฎ๐˜ƒ๐—ถ๐—ผ๐˜‚๐—ฟ ๐—”๐—ป๐—ฎ๐—น๐˜†๐˜€๐—ถ๐˜€ ๐—ณ๐—ผ๐—ฟ ๐—”๐—ฟ๐˜๐—ถ๐—ณ๐—ถ๐—ฐ๐—ถ๐—ฎ๐—น ๐— ๐—ฒ๐—ฑ๐—ถ๐—ฎ๐˜๐—ถ๐—ผ๐—ป", has been accepted as a Grand Challenge at ACM Multimedia 2025!๐Ÿš€

"The goal of this multi-year challenge is to contribute to realising the vision of autonomous artificial mediators by measurable advances in key conversational behaviour sensing and analysis tasks."

For more details, visit https://www.multimediate-challenge.org/.

#MM25

MultiMediate:Multi-modal Group Behaviour Analysis for Artificial Mediation

Grand Challenge at ACM MMโ€™25

๐ŸŒŸ Exciting News! ๐ŸŒŸ Our groupโ€™s paper ๐—ฉยฒ๐——๐—ถ๐—ฎ๐—น: ๐—จ๐—ป๐—ถ๐—ณ๐—ถ๐—ฐ๐—ฎ๐˜๐—ถ๐—ผ๐—ป ๐—ผ๐—ณ ๐—ฉ๐—ถ๐—ฑ๐—ฒ๐—ผ ๐—ฎ๐—ป๐—ฑ ๐—ฉ๐—ถ๐˜€๐˜‚๐—ฎ๐—น ๐——๐—ถ๐—ฎ๐—น๐—ผ๐—ด ๐˜ƒ๐—ถ๐—ฎ ๐— ๐˜‚๐—น๐˜๐—ถ๐—บ๐—ผ๐—ฑ๐—ฎ๐—น ๐—˜๐˜…๐—ฝ๐—ฒ๐—ฟ๐˜๐˜€ has been accepted at CVPR 2025! ๐ŸŽ‰๐Ÿ“š

VยฒDial is a novel model specifically designed to handle both image and video input data for multimodal conversational tasks. Extensive evaluations on AVSD and VisDial datasets show that VยฒDial achieves new state-of-the-art results across multiple benchmarks.

Congratulations to the authors. ๐Ÿ™Œ

#CVPR2025 #ComputerVision #DeepLearning

๐Ÿ“ข New Paper Alert! ๐Ÿ“ข

We're thrilled to announce that our paper, HAIFAI: Human-AI Interaction for Mental Face Reconstruction, has been accepted by ACM Transactions on Interactive Intelligent Systems (TiiS).

Congratulations to the authors!

You can check out the preprint on arXiv https://arxiv.org/abs/2412.06323v1 and stay tuned for the camera-ready version on our website https://collaborative-ai.org/.

HAIFAI: Human-AI Collaboration for Mental Face Reconstruction

We present HAIFAI - a novel collaborative human-AI system to tackle the challenging task of reconstructing a visual representation of a face that exists only in a person's mind. Users iteratively rank images presented by the AI system based on their resemblance to a mental image. These rankings, in turn, allow the system to extract relevant image features, fuse them into a unified feature vector, and use a generative model to reconstruct the mental image. We also propose an extension called HAIFAI-X that allows users to manually refine and further improve the reconstruction using an easy-to-use slider interface. To avoid the need for tedious human data collection for model training, we introduce a computational user model of human ranking behaviour. For this, we collected a small face ranking dataset through an online crowd-sourcing study containing data from 275 participants. We evaluate HAIFAI and HAIFAI-X in a 12-participant user study and show that HAIFAI outperforms the previous state of the art regarding reconstruction quality, usability, perceived workload, and reconstruction speed. HAIFAI-X achieves even better reconstruction quality at the cost of reduced usability, perceived workload, and increased reconstruction time. We further validate the reconstructions in a subsequent face ranking study with 18 participants and show that HAIFAI-X achieves a new state-of-the-art identification rate of 60.6%. These findings represent a significant advancement towards developing new collaborative intelligent systems capable of reliably and effortlessly reconstructing a user's mental image.

arXiv.org

๐ŸŽ‰ Exciting News! ๐ŸŽ‰

Our group has two papers (conditionally) accepted to #CHI2025.

1๏ธโƒฃ SummAct: Uncovering User Intentions Through Interactive Behaviour Summarisation

2๏ธโƒฃ How People Read Charts: A Model of Task-driven Eye Movement Control

Stay tuned for more details about the papers at: https://collaborative-ai.org/news/2025/01/two-paper-accepted-at-chi/

Collaborative Artificial Intelligence

Our group conducts fundamental research towards collaborative artificial intelligence (CAI) at the intersection of multimodal machine learning, computational cognitive modelling, computer vision, and human-machine interaction.

๐ŸŽ‰We are thrilled to share that our lab director Prof. @abulling has joined the editorial board of the IEEE Transactions on Visualization and Computer Graphics (TVCG) as of January 1, 2025.

Congratulations, Prof. Bulling! ๐ŸŒŸ

๐Ÿš€ Exciting News! ๐Ÿš€

Our lab is a part of the ICRA 2025 workshop on Nonverbal Cues for Human-Robot Cooperative Intelligence! ๐Ÿค–โœจ

This workshop is organized by an incredible team, including our own alumnus Prof. Xucong Zhang and lab director Prof. Andreas Bulling @abulling.

Stay tuned for more updates from this exciting event! ๐ŸŒŸ

https://nocworkshop.github.io/2025/#

#ICRA2025

ICRA 2025: The 2nd Workshop on Nonverbal Cues for Human-Robot Cooperative Intelligence

We are excited to announce that the funding proposal "Privacy-preserving Eye Tracking" has been approved by the German Research Foundation (DFG)! ๐ŸŽ‰

It is a joint proposal with Prof. Ralf Kรผsters at the Institute of Information Security, and will fund one PostDoc and one PhD position for three years.

The goal of this project is to advance privacy-preserving eye tracking along the whole eye tracking data processing pipeline.