Paper
Who's Sense is This? Possibility for Impacting Human Insights in AI-assisted Sensemaking
Authors
Zhuoyi Cheng, Steven Houben
Abstract
Sensemaking is an important preceding step for activities like consensus building and decision-making. When groups of people make sense of large amounts of information, their understanding gradually evolves from vague to clear. During this process when reaching a conclusion is still premature, if people are presented with others' insights, they may be directed to focus on that specific perspective without adequate verification. We argue that similar phenomena may also exist in AI-assisted sensemaking, in which AI will usually be the one that presents insight prematurely when users' understandings are still vague and ill-formed. In this paper, we raised three questions that are worth deliberation before exploiting AI to assist in collaborative sensemaking in practice, and discussed possible reasons that may lead users to opt for insights from AI.
Metadata
Related papers
Vibe Coding XR: Accelerating AI + XR Prototyping with XR Blocks and Gemini
Ruofei Du, Benjamin Hersh, David Li, Nels Numan, Xun Qian, Yanhe Chen, Zhongy... • 2026-03-25
Comparing Developer and LLM Biases in Code Evaluation
Aditya Mittal, Ryan Shar, Zichu Wu, Shyam Agarwal, Tongshuang Wu, Chris Donah... • 2026-03-25
The Stochastic Gap: A Markovian Framework for Pre-Deployment Reliability and Oversight-Cost Auditing in Agentic Artificial Intelligence
Biplab Pal, Santanu Bhattacharya • 2026-03-25
Retrieval Improvements Do Not Guarantee Better Answers: A Study of RAG for AI Policy QA
Saahil Mathur, Ryan David Rittner, Vedant Ajit Thakur, Daniel Stuart Schiff, ... • 2026-03-25
MARCH: Multi-Agent Reinforced Self-Check for LLM Hallucination
Zhuo Li, Yupeng Zhang, Pengyu Cheng, Jiajun Song, Mengyu Zhou, Hao Li, Shujie... • 2026-03-25
Raw Data (Debug)
{
"raw_xml": "<entry>\n <id>http://arxiv.org/abs/2603.17643v1</id>\n <title>Who's Sense is This? Possibility for Impacting Human Insights in AI-assisted Sensemaking</title>\n <updated>2026-03-18T12:04:54Z</updated>\n <link href='https://arxiv.org/abs/2603.17643v1' rel='alternate' type='text/html'/>\n <link href='https://arxiv.org/pdf/2603.17643v1' rel='related' title='pdf' type='application/pdf'/>\n <summary>Sensemaking is an important preceding step for activities like consensus building and decision-making. When groups of people make sense of large amounts of information, their understanding gradually evolves from vague to clear. During this process when reaching a conclusion is still premature, if people are presented with others' insights, they may be directed to focus on that specific perspective without adequate verification. We argue that similar phenomena may also exist in AI-assisted sensemaking, in which AI will usually be the one that presents insight prematurely when users' understandings are still vague and ill-formed. In this paper, we raised three questions that are worth deliberation before exploiting AI to assist in collaborative sensemaking in practice, and discussed possible reasons that may lead users to opt for insights from AI.</summary>\n <category scheme='http://arxiv.org/schemas/atom' term='cs.HC'/>\n <published>2026-03-18T12:04:54Z</published>\n <arxiv:comment>Accepted by CHI 26 Workshop on Sensemaking and AI 2026: Uses, Behaviors, Design, and Recommendations</arxiv:comment>\n <arxiv:primary_category term='cs.HC'/>\n <author>\n <name>Zhuoyi Cheng</name>\n </author>\n <author>\n <name>Steven Houben</name>\n </author>\n </entry>"
}