Loading Session...

Resource I: Interactive and Conversational Search

Session Information

  • WildClaims: Conversational Information Access in theWild(Chat)
  • LISP - A Rich Interaction Dataset and Loggable Interactive Search Platform
  • UserSimCRS v2: Simulation-Based Evaluation for Conversational Recommender Systems
  • Sim4IA-Bench: A User Simulation Benchmark Suite for Next Query and Utterance Prediction
  • Beyond the Click: A Framework for Inferring Cognitive Traces in Search
Mar 31, 2026 14:30 - 16:00(Europe/Amsterdam)
Venue : Chemie
20260331T1430 20260331T1600 Europe/Amsterdam Resource I: Interactive and Conversational Search WildClaims: Conversational Information Access in theWild(Chat)LISP - A Rich Interaction Dataset and Loggable Interactive Search PlatformUserSimCRS v2: Simulation-Based Evaluation for Conversational Recommender SystemsSim4IA-Bench: A User Simulation Benchmark Suite for Next Query and Utterance PredictionBeyond the Click: A Framework for Inferring Cognitive Traces in Search Chemie ECIR2026 conference-secretariat@blueboxevents.nl

Sub Sessions

Beyond the Click: A Framework for Inferring Cognitive Traces in Search

ResourceUser aspects in IRResource 02:30 PM - 04:00 PM (Europe/Amsterdam) 2026/03/31 12:30:00 UTC - 2026/03/31 14:00:00 UTC
User simulators are essential for evaluating search systems, but they primarily copy user actions without understanding the underlying thought process. This gap exists because large-scale interaction logs record what users do, but not what they might be thinking or feeling, such as confusion or satisfaction. To solve this problem, we present a new framework that computationally infers cognitive traces from behavioral data. Our method uses a multi-agent language model system, grounded in Information Foraging Theory and calibrated by human experts, to annotate user actions with their likely cognitive state. To show the value of these traces, we demonstrate that they significantly improve a model's ability to predict when a user will abandon a search task. We release a collection of annotations for several public datasets, including AOL and Stack Overflow, and an open-source tool that allows researchers to apply our method to their own data. This work provides the tools and data needed to build more human-like user simulators and to assess retrieval systems on user-oriented dimensions of performance.
Presenters
SZ
Saber Zerhoudi
Postdoctoral Researcher, University Of Passau
Co-Authors
MG
Michael Granitzer
University Of Passau

Sim4IA-Bench: A User Simulation Benchmark Suite for NextQuery and Utterance Prediction

ResourceEvaluation research User aspects in IR 02:30 PM - 04:00 PM (Europe/Amsterdam) 2026/03/31 12:30:00 UTC - 2026/03/31 14:00:00 UTC
Validating user simulation is a difficult task due to the lack of established measures and benchmarks, which makes it challenging to assess whether a simulator accurately reflects real user behavior. As part of the Sim4IA Micro-Shared Task at the Sim4IA Workshop, SIGIR 2025, we present Sim4IA-Bench, a simulation benchmark suit for the prediction of the next queries and utterances, the first of its kind in the IR community. Our dataset as part of the suite comprises 160 real-world search sessions from the CORE search engine. For 70 of these sessions, up to 62 simulator runs are available, divided into Task A and Task B, in which different approaches predicted users¡¯ next search queries or utterances. Sim4IA-Bench provides a basis for evaluating and comparing user simulation approaches and for developing new measures of simulator validity. Although modest in size, the suite represents the first publicly available benchmark that links real search sessions with simulated next-query predictions. In addition to serving as a testbed for next query prediction, it also enables exploratory studies on query reformulation behavior, intent drift, and interaction-aware retrieval evaluation. We also introduce a new measure for evaluating next-query predictions in this task. By making the suite publicly available, we aim to promote reproducible research and stimulate further work on realistic and explainable user simulation for information access: https://github.com/irgroup/Sim4IA-Bench.
Presenters
AK
Andreas Konstantin Kruff
PhD Student, TH Köln
Co-Authors
CK
Christin Katharina Kreutz
TH Mittelhessen
TB
Timo Breuer
TH K?ln
PS
Philipp Schaer
Professor, TH Köln
KB
Krisztian Balog
Professor, University Of Stavanger

UserSimCRS v2: Simulation-Based Evaluation for Conversational Recommender Systems

ResourceResource 02:30 PM - 04:00 PM (Europe/Amsterdam) 2026/03/31 12:30:00 UTC - 2026/03/31 14:00:00 UTC
Resources for simulation-based evaluation of conversational recommender systems (CRSs) are scarce. The UserSimCRS toolkit was introduced to address this gap. In this work, we present UserSimCRS v2, a significant upgrade aligning the toolkit with state-of-the-art research. Key extensions include an enhanced agenda-based user simulator, introduction of large language model-based simulators, integration for a wider range of CRSs and datasets, and new LLM-as-a-judge evaluation utilities. We demonstrate these extensions in a case study.
Presenters
NB
Nolwenn Bernard
TH Köln
Co-Authors
KB
Krisztian Balog
Professor, University Of Stavanger

LISP - A Rich Interaction Dataset and Loggable InteractiveSearch Platform

ResourceUser aspects in IR 02:30 PM - 04:00 PM (Europe/Amsterdam) 2026/03/31 12:30:00 UTC - 2026/03/31 14:00:00 UTC
We present a reusable dataset and accompanying infrastructure for studying human search behavior in Interactive Information Retrieval (IIR). The dataset combines detailed interaction logs from 61 participants (122 sessions) with user characteristics, including perceptual speed, topic-specific interest, search expertise, and demographic information. To facilitate reproducibility and reuse, we provide a fully documented study setup, a web-based perceptual speed test, and a framework for conducting similar user studies. Our work allows researchers to investigate individual and contextual factors affecting search behavior, and to develop or validate user simulators that account for such variability. We illustrate the dataset¡¯s potential through an illustrative analysis and release all resources as open-access, supporting reproducible research and resource sharing in the IIR community.
Presenters
JF
Jana Friese
PhD Student, University Of Duisburg-Essen
Co-Authors
AK
Andreas Konstantin Kruff
PhD Student, TH Köln
PS
Philipp Schaer
Professor, TH Köln
NF
Nicola Ferro
Full Professor, University Of Padova

WildClaims: Conversational Information Access in the Wild(Chat)

ResourceConversational search and recommender systemsResource 02:30 PM - 04:00 PM (Europe/Amsterdam) 2026/03/31 12:30:00 UTC - 2026/03/31 14:00:00 UTC
The rapid advancement of Large Language Models (LLMs) has transformed conversational systems into practical tools used by millions. However, the nature and necessity of information retrieval in real-world conversations remain largely unexplored, as research has focused predominantly on traditional, explicit information access conversations. The central question is: What does real-world conversational information access look like? To this end, we first conduct an observational study on the WildChat dataset, large-scale user-ChatGPT conversations, finding that users' access to information occurs implicitly as check-worthy factual assertions made by the system, even when the conversation's primary intent is non-informational, such as creative writing. To enable the systematic study of this phenomenon, we release the WildClaims dataset, a novel resource consisting of 121,905 extracted factual claims from 7,587 utterances in 3,000 WildChat conversations, each annotated for check-worthiness. Our preliminary analysis of this resource reveals that conservatively 18% to 51% of conversations contain check-worthy assertions, depending on the methods employed, and less conservatively, as many as 76% may contain such assertions. This high prevalence underscores the importance of moving beyond the traditional understanding of explicit information access, to address the implicit information access that arises in real-world user-system conversations.
Presenters
HJ
Hideaki Joko
Radboud University
Co-Authors Shakiba Amirshahi
Master Student, University Of Waterloo
CC
Charles Clarke
Professor, University Of Waterloo
FH
Faegheh Hasibi
Radboud University
176 visits

Session Participants

User Online
Session speakers, moderators & attendees
Radboud University
PhD student
,
University Of Duisburg-Essen
PhD Student
,
TH Köln
Postdoctoral researcher
,
University Of Passau
PhD Student
,
Friedrich-Schiller-Universität Jena
 Sérgio Nunes
University of Porto | INESC TEC
24 attendees saved this session

Session Chat

Live Chat
Chat with participants attending this session

Questions & Answers

Answered
Submit questions for the presenters

Session Polls

Active
Participate in live polls

Need Help?

Technical Issues?

If you're experiencing playback problems, try adjusting the quality or refreshing the page.

Questions for Speakers?

Use the Q&A tab to submit questions that may be addressed in follow-up sessions.