Publications
[Google Scholar] [ResearchGate] [DBLP] [ACM] [ORCID] [RMIT]
2025
Charles L. A. Clarke, Paul Kantor, Adam Roegiest, Ian Soboroff, Johanne R. Trippas, and Zhaochun Ren. The Second Search Futures Workshop. In Proceedings of the European Conference on Information Retrieval (ECIR’25), ECIR ’25, pages 1–4, 2025.
[PDF] [DOI] [Abstract] [Cite]
Abstract
Cite
@inproceedings{clarke2025second,
title={{The Second Search Futures Workshop}},
author={Charles L. A. Clarke and Paul Kantor and Adam Roegiest and Ian Soboroff and Johanne~R. Trippas and Zhaochun Ren},
booktitle={Proceedings of the European Conference on Information Retrieval (ECIR'25)},
pages = {1--4},
year={2025},
series = {ECIR '25}
}
2024
Johanne R. Trippas, Sara F. D. Al Lawati, Joel Mackenzie and Luke Gallagher. What do Users Really Ask Large Language Models? An Initial Log Analysis of Google Bard Interactions in the Wild. In Proceedings of the 47th International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR’24), 2024.
[PDF] [BIDD-1k Dataset] [DOI] [Abstract] [Cite] [Poster]
Abstract
Advancements in large language models (LLMs) have changed information retrieval, offering users a more personalised and natural search experience with technologies like OpenAI ChatGPT, Google Bard (Gemini), or Microsoft Copilot. Despite these advancements, research into user tasks and information needs remains scarce. This preliminary work analyses a Google Bard prompt log with 15,023 interactions called the Bard Intelligence and Dialogue Dataset (BIDD), providing an understanding akin to query log analyses. We show that Google Bard prompts are often verbose and structured, encapsulating a broader range of information needs and imperative (e.g., directive) tasks distinct from traditional search queries. We show that LLMs can support users in tasks beyond the three main types based on user intent: informational, navigational, and transactional. Our findings emphasise the versatile application of LLMs across content creation, LLM writing style preferences, and information extraction. We document diverse user interaction styles, showcasing the adaptability of users to LLM capabilities.
Cite
@inproceedings{trippas2024what,
author = {Trippas, Johanne R and Al Lawati, Sara Fahad Dawood and Mackenzie, Joel and Gallagher, Luke},
booktitle = {Proceedings of the 47th International ACM SIGIR Conference on Research and Development in Information Retrieval},
series = {SIGIR '24},
title = {What do Users Really Ask Large Language Models? An Initial Log Analysis of Google Bard Interactions in the Wild},
year = {2024},
doi = {10.1145/3626772.3657914},
address = {New York, NY, USA},
location = {Washington, DC, USA},
publisher = {ACM}
}
Johanne R. Trippas, Luke Gallagher, Joel Mackenzie. Re-evaluating the Command-and-Control Paradigm in Conversational Search Interactions. In Proceedings of the 33rd ACM International Conference on Information and Knowledge Management (CIKM ’24), 2024.
[PDF] [Dataset] [DOI] [Abstract] [Cite]
Abstract
Conversational assistants are becoming prevalent among the wider population due to their simplicity and increasing utility. However, the shortcomings of these tools are as renowned as their benefits. In this work, we present a “first look” at an extensive collection of conversational queries, aiming to identify limitations and improvement opportunities specifically related to information access (i.e., search interactions). We explore over 600,000 Google Assistant interactions from 173 unique users, examining usage trends and the resulting deficiencies and strengths of these assistants. We aim to provide a balanced assessment, highlighting the assistant’s shortcomings in supporting users and delivering relevant information to user needs and areas where it demonstrates a reasonable response to user inputs. Our analysis shows that, although most users conduct information-seeking tasks, there is little evidence of complex information-seeking behaviour, with most interactions consisting of simple, imperative instructions. Finally, we find that conversational devices allow users to benefit from increased naturalistic interactions and the ability to apply acquired information in situ, a novel observation for conversational information seeking.
Cite
@inproceedings{trippas2024reevaluating,
author = {Trippas, Johanne R and Gallagher, Luke and Mackenzie, Joel},
booktitle = {Proceedings of the 33rd ACM International Conference on Information and Knowledge Management},
series = {CIKM '24},
title = {Re-evaluating the Command-and-Control Paradigm in Conversational Search Interactions},
year = {2024},
doi = {doi.org/10.1145/3627673.3679588},
address = {New York, NY, USA},
location = {Boise, ID, USA},
publisher = {ACM}
}
Weronika Łajewska, Damiano Spina, Johanne R. Trippas, Krisztian Balog. Explainability for Transparent Conversational Information-Seeking. In Proceedings of the 47th International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR’24), 2024.
[PDF] [Dataset] [DOI] [Abstract] [Cite]
Abstract
The increasing reliance on digital information necessitates advancements in conversational search systems, particularly in terms of information transparency. While prior research in conversational information-seeking has concentrated on improving retrieval techniques, the challenge remains in generating responses useful from a user perspective. This study explores different methods of explaining the responses, hypothesizing that transparency about the source of the information, system confidence, and limitations can enhance users’ ability to objectively assess the response. By exploring transparency across explanation type, quality, and presentation mode, this research aims to bridge the gap between system-generated responses and responses verifiable by the user. We design a user study to answer questions concerning the impact of (1) the quality of explanations enhancing the response on its usefulness and (2) ways of presenting explanations to users. The analysis of the collected data reveals lower user ratings for noisy explanations, although these scores seem insensitive to the quality of the response. Inconclusive results on the explanations presentation format suggest that it may not be a critical factor in this setting.
Cite
@inproceedings{lajewska2024explainability,
author = {Łajewska, Weronika and Spina, Damiano and Trippas, Johanne R. and Balog, Krisztian},
booktitle = {Proceedings of the 47th International ACM SIGIR Conference on Research and Development in Information Retrieval},
series = {SIGIR '24},
title = {Explainability for Transparent Conversational Information-Seeking},
year = {2024},
doi = {10.1145/3626772.3657768},
address = {New York, NY, USA},
location = {Washington, DC, USA},
publisher = {ACM}
}
Vahid Sadiri Javadi, Johanne R. Trippas, Lucie Flek. Unveiling Information Through Narrative In Conversational Information Seeking. In Proceedings of the ACM International Conference on Conversational User Interfaces (CUI 2024), 2024.
[PDF] [DOI] [Abstract] [Cite]
Abstract
Searching through conversational interactions has been emphasized as the next frontier. Nowadays, conversational agents can generate natural language responses, transforming how we search for information. A key challenge in conversational information-seeking is how these agents present information: should they only reflect facts, cater to human cognitive preferences, or strike a balance between them? These challenges raise questions about aligning conversational agents with human cognitive processes. Our position paper emphasizes the role of narrative in addressing these questions. We explore how narratives influence human comprehension and propose a framework for optimal conversational narratives. These narratives aim to enhance interaction between humans and conversational agents in explanatory information-seeking scenarios.
Cite
@inproceedings{sadiri2024unveiling,
title ={Unveiling Information Through Narrative In Conversational Information Seeking},
Author = {Sadiri Javadi, Vahid and Trippas, Johanne R. and Lucie Flek},
Booktitle = {Proceedings of the ACM International Conference on Conversational User Interfaces (CUI 2024)},
Year = {2024},
numpages = {6},
pages = {1--6},
doi={10.1145/3640794.3665884},
address = {New York, NY, USA},
location = {Washington, DC, USA},
publisher = {ACM}
}
Leila Tavakoli, Johanne R. Trippas, Hamed Zamani, Falk Scholer, and Mark Sanderson. Online and Offline Evaluation in Search Clarification. In ACM Transactions on Information Systems (ACM TOIS), 2024.
[PDF] [DOI] [Abstract] [Cite]
Abstract
The effectiveness of clarification question models in engaging users within search systems is currently constrained, casting doubt on their overall usefulness. To improve the performance of these models, it is crucial to employ assessment approaches that encompass both real-time feedback from users (online evaluation) and the characteristics of clarification questions evaluated through human assessment (offline evaluation). However, the relationship between online and offline evaluations has been debated in information retrieval. This study aims to investigate how this discordance holds in search clarification. We use user engagement as ground truth and employ several offline labels to investigate to what extent the offline ranked lists of clarification resemble the ideal ranked lists based on online user engagement. Contrary to the current understanding that offline evaluations fall short of supporting online evaluations, we indicate that when identifying the most engaging clarification questions from the user’s perspective, online and offline evaluations correspond with each other. We show that the query length does not influence the relationship between online and offline evaluations, and reducing uncertainty in online evaluation strengthens this relationship. We illustrate that an engaging clarification needs to excel from multiple perspectives, and SERP quality and characteristics of the clarification are equally important. We also investigate if human labels can enhance the performance of Large Language Models (LLMs) and Learning-to-Rank (LTR) models in identifying the most engaging clarification questions from the user’s perspective by incorporating offline evaluations as input features. Our results indicate that Learning-to-Rank models do not perform better than individual offline labels. However, GPT, an LLM, emerges as the standout performer, surpassing all Learning-to-Rank models and offline labels.
Cite
@inproceedings{tavakoli2024online,
title ={Online and Offline Evaluation in Search Clarification},
author = {Tavakoli, Leila and Trippas, Johanne R. and Zamani, Hamed and Scholer, Falk and Sanderson, Mark},
journal = {ACM Transactions on Information Systems (TOIS)},
year = {2024},
numpages = {28},
doi = {10.1145/3681786},
address = {New York, NY, USA},
location = {Washington, DC, USA},
publisher = {ACM}
}
Johanne R. Trippas, Damiano Spina, and Falk Scholer. Adapting Generative Information Retrieval Systems to Users, Tasks, and Scenarios. In Ryen W. White and Chirag Shah, editors, Information Access in the Era of Generative AI. Springer Nature, 2024.
[PDF] [DOI] [Abstract] [Cite]
Abstract
Generative Information Retrieval (GenIR) signifies an advancement in Information Retrieval (IR). GenIR encourages more sophisticated, conversational responses to search queries by integrating generative models and chat-like interfaces. However, this approach retains core principles of traditional IR and conversational information seeking, illustrating its capacity to augment current IR frameworks. In this chapter, we propose that introducing GenIR enhances traditional information retrieval tasks and expands their scope. This allows systems to manage more complex queries, including generative, critiquing, and extractive tasks. These advancements surpass traditional systems, handling queries with greater depth and flexibility. This sometimes speculative chapter suggests Generative Information Access (GenIA), a term that more accurately encapsulates the widened scope and enhanced functionalities of GenIR, particularly in how this relates to tasks. By investigating the impact of GenIR, this discussion aims to reiterate that generative research should not abandon traditional interactive information retrieval research but rather incorporate it into future research and development efforts.
Cite
@incollection{trippas2024adapting,
address = {Cham, Switzerland},
author = {Trippas, Johanne R. and Spina, Damiano and Scholer, Falk},
booktitle = {Information Access in the Era of Generative AI},
editor = {White, Ryen W. and Shah, Chirag},
publisher = {Springer Nature Switzerland AG},
title = {Adapting Generative Information Retrieval Systems to Users, Tasks, and Scenarios},
year = {2024}
}
S. P. Cherumanal, L. Tian, F. M. Abushaqra, A. F. M. de Paula, K. Ji, H. Ali, D. Hettiachchi, J. R. Trippas, F. Scholer, and D. Spina. Walert: Putting Conversational Search Knowledge into Action by Building and Evaluating a Large Language Model-Powered Chatbot. In Proceedings of the ACM Conference on Information Interaction and Retrieval (CHIIR’24), pages 1–10, 2024.
[PDF] [Github] [DOI] [Abstract] [Cite]
Abstract
Creating and deploying customized applications is crucial for operational success and enriching user experiences in the rapidly evolving modern business world. A prominent facet of modern user experiences is the integration of chatbots or voice assistants. The rapid evolution of Large Language Models (LLMs) has provided a powerful tool to build conversational applications. We present Walert, a customized LLM-based conversational agent able to answer frequently asked questions about computer science degrees and programs at RMIT University. Our demo aims to showcase how conversational information-seeking researchers can effectively communicate the benefits of using best practices to stakeholders interested in developing and deploying LLM-based chatbots. These practices are well-known in our community but often overlooked by practitioners who may not have access to this knowledge. The methodology and resources used in this demo serve as a bridge to facilitate knowledge transfer from experts, address industry professionals’ practical needs, and foster a collaborative environment. The data and code of the demo are available at https://github.com/rmit-ir/walert.
Cite
@inproceedings{pathiyan2024walert,
author = {Sachin Pathiyan Cherumanal and Lin Tian and Futoon M.~Abushaqra and Angel F.~ Magnossao de Paula and Kaixin Ji and Halil Ali and Danula Hettiachchi and Johanne~R. Trippas and Falk Scholer and Damiano Spina},
booktitle = {Proceedings of the ACM Conference on Information Interaction and Retrieval},
series = {CHIIR '24},
title = {{Walert: Putting Conversational Search Knowledge into Action by Building and Evaluating a Large Language Model-Powered Chatbot}},
year = {2024},
doi = {10.1145/3627508.3638309},
address = {New York, NY, USA},
location = {Washington, DC, USA},
publisher = {ACM}
}
Johanne R. Trippas and David Maxwell. PhD Candidacy: A Tutorial on Overcoming Challenges and Achieving Success. In Proceedings of the European Conference on Information Retrieval (ECIR’24), ECIR ’24, pages 1–4, 2024.
[PDF]
Adam Roegiest and Johanne R. Trippas. UnExplored FrontCHIIRs: A Workshop Exploring Future Directions for Information Access. In Proceedings of the ACM Conference on Information Interaction and Retrieval (CHIIR’24), CHIIR ’24, pages 1–4, 2024.
[PDF] [DOI] [Abstract] [Cite]
Abstract
With the rise and growing prevalence of generative models, particularly multi-modal ones, it is an opportune time to explore beyond existing interactive information retrieval research trends. Indeed, it is essential to determine new avenues to explore how users interact with these models as well as revisit existing avenues that can be embellished with new technology. In this session, we aim to create a venue to workshop ideas that explore the future of search experiences and user interactions with information in a collaborative, low-pressure environment. This UnExplored FrontCHIIRs workshop enables participants to form a sub-community within CHIIR to facilitate further development of the proposed ideas and allow deeper collaborative problem-solving than just presenting late-breaking work.
Cite
@inproceedings{roegiest2024unexplored,
title={UnExplored FrontCHIIRs: A Workshop Exploring Future Directions for Information Access},
author={Roegiest, Adam and Trippas, Johanne},
booktitle={Proceedings of the 2024 Conference on Human Information Interaction and Retrieval},
pages={436--437},
year={2024},
doi = {10.1145/3627508.3638302},
address = {New York, NY, USA},
location = {Washington, DC, USA},
publisher = {ACM}
}
Leif Azzopardi, Charles L. A. Clarke, Paul Kantor, Bhaskar Mitra, Johanne R. Trippas, and Zhaochun Ren. Report on The Search Futures Workshop at ECIR 2024. In SIGIR Forum, 58(1), 2024.
[PDF] [Abstract] [Cite]
Abstract
The First Search Futures Workshop, in conjunction with the Fourty-sixth European Conference on Information Retrieval (ECIR) 2024, looked into the future of search to ask questions such as:
-How can we harness the power of generative AI to enhance, improve and re-imagine Information Retrieval (IR)?
-What are the principles and fundamental rights that the field of Information Retrieval should strive to uphold?
-How can we build trustworthy IR systems in light of Large Language Models and their ability to generate content at super human speeds?
-What new applications and affordances does generative AI offer and enable, and can we go back to the future, and do what we only dreamed of previously?
The workshop started with seventeen lightning talks from a diverse set speakers. Instead of conventional paper presentations, the lightning talks provided a rapid and concise overview of ideas, allowing speakers to share critical points or novel concepts quickly. This format was designed to encourage discussion and introduce a wide range of topics within a short period, thereby maximising the exchange of ideas and ensuring that participants could gain insights into various future search areas without the deep dive typically required in longer presentations. This report, co-authored by the workshop’s organisers and its participants, summarises the talks and discussions. This report aims to provide the broader IR community with the insights and ideas discussed and debated during the workshop – and to provide a platform for future discussion.
Cite
@article{azzopardi2024report,
title={Report on The Search Futures Workshop at ECIR 2024},
author={Leif Azzopardi and Charles L. A. Clarke and Paul Kantor and Bhaskar Mitra and Johanne R. Trippas and Zhaochun Ren},
year = {2024},
issue_date = {June 2024},
publisher = {Association for Computing Machinery},
address = {New York, NY, USA},
volume = {58},
number = {1},
journal = {SIGIR Forum},
numpages = {41}
}
Kaixin Ji, Sashin Pathiyan Cherumanal, Johanne R. Trippas, Danula Hettiachchi, Flora D. Salim, Falk Scholer, and Damiano Spina. Towards detecting and mitigating cognitive bias in spoken conversational search. In Adjunct Publication of the 26th International Conference on Human-Computer Interaction with Mobile Devices and Services, MobileHCI ’24, 2024.
[PDF] [DOI] [Abstract] [Cite]
Abstract
Spoken Conversational Search (SCS) poses unique challenges in understanding user-system interactions due to the absence of visual cues, and the complexity of less structured dialogue. Tackling the impacts of cognitive bias in today’s information-rich online environment, especially when SCS becomes more prevalent, this paper integrates insights from information science, psychology, cognitive science, and wearable sensor technology to explore potential opportunities and challenges in studying cognitive biases in SCS. It then outlines a framework for experimental designs with various experiment setups to multimodal instruments. It also analyzes data from an existing dataset as a preliminary example to demonstrate the potential of this framework and discuss its implications for future research. In the end, it discusses the challenges and ethical considerations associated with implementing this approach. This work aims to provoke new directions and discussion in the community and enhance understanding of cognitive biases in Spoken Conversational Search.
Cite
@inproceedings{ji2024towards,
author = {Ji, Kaixin and Pathiyan Cherumanal, Sachin and Johanne R. Trippas and Hettiachchi, Danula
and Salim, Flora D. and Scholer, Falk and Spina, Damiano},
booktitle = {Adjunct Publication of the 26th International Conference on Human-Computer Interaction with Mobile Devices and Services},
doi = {10.1145/3640471.3680245},
series = {MobileHCI '24},
title = {Towards Detecting and Mitigating Cognitive Bias in Spoken Conversational Search},
year = {2024}
}
Sashin Pathiyan Cherumanal, Falk Scholer, Johanne R. Trippas, and Damiano Spina. Towards Investigating Biases in Spoken Conversational Search. In ACM International Conference on Multimodal Interaction, ICMI ’24, 2024.
[PDF] [DOI] [Abstract] [Cite]
Abstract
Voice-based systems like Amazon Alexa, Google Assistant, and Apple Siri, along with the growing popularity of OpenAI’s Chat-GPT and Microsoft’s Copilot, serve diverse populations, including visually impaired and low-literacy communities. This reflects a shift in user expectations from traditional search to more interactive question-answering models. However, presenting information effectively in voice-only channels remains challenging due to their linear nature. This limitation can impact the presentation of complex queries involving controversial topics with multiple perspectives. Failing to present diverse viewpoints may perpetuate or introduce biases and affect user attitudes. Balancing information load and addressing biases is crucial in designing a fair and effective voice-based system. To address this, we (i) review how biases and user attitude changes have been studied in screen-based web search, (ii) address challenges in studying these changes in voice-based settings like SCS, (iii) outline research questions, and (iv) propose an experimental setup with variables, data, and instruments to explore biases in a voice-based setting like Spoken Conversational Search.
Cite
@inproceedings{pathiyan2024towards,
author = {Pathiyan Cherumanal, Sachin and Scholer, Falk and Johanne R. Trippas and Spina, Damiano},
booktitle = {ACM International Conference on Multimodal Interaction},
doi = {10.1145/3686215.3690156},
series = {ICMI '24},
title = {{Towards Investigating Biases in Spoken Conversational Search}},
numpages = {6},
year = {2024}
}
Weronika Łajewska, Krisztian Balog, Damiano Spina, Johanne R. Trippas. Can Users Detect Biases or Factual Errors in Generated Responses in Conversational Information-Seeking?. In Annual International ACM SIGIR Conference on Research and Development in Information Retrieval in the Asia Pacific Region (SIGIR-AP’24), 2024.
[PDF] [Dataset] [DOI] [Abstract] [Cite]
Abstract
Cite
@inproceedings{lajewska2024explainability,
author = {Łajewska, Weronika and Balog, Krisztian and Spina, Damiano and Trippas, Johanne R.},
booktitle = {Annual International ACM SIGIR Conference on Research and Development in Information Retrieval in the Asia Pacific Region},
series = {SIGIR-AP '24},
title = {{Can Users Detect Biases or Factual Errors in Generated Responses in Conversational Information-Seeking?}},
year = {2024},
doi = {},
address = {New York, NY, USA},
location = {Tokyo, Japan},
publisher = {ACM}
}
Vahid Sadiri Javadi, Johanne R. Trippas, Lucie Flek. Can Stories Help LLMs Reason? Curating Information Space Through Narrative. In The First Workshop on System-2 Reasoning at Scale, NeurIPS’24, 2024.
[PDF] [DOI] [Abstract] [Cite]
Abstract
Cite
@inproceedings{javadi2024can,
title={Can Stories Help {LLM}s Reason? Curating Information Space Through Narrative},
author={Vahid Sadiri Javadi and Johanne Trippas and Lucie Flek},
booktitle={The First Workshop on System-2 Reasoning at Scale, NeurIPS'24},
year={2024}
}
Leila Tavakoli, Giovanni Castiglia, Federica Calò, Yashar Deldjoo, Hamed Zamani, and Johanne R. Trippas. Understanding Modality Preferences in Search Clarification. In 1st Workshop on Multimodal Search and Recommendations (CIKM MMSR ’24), 2024.
[PDF] [Abstract] [Cite]
Abstract
This study is the first attempt to explore the impact of clarification question modality on user preference in search engines. We introduce the multi-modal search clarification dataset, MIMICS-MM, containing clarification questions with associated expert-collected and model-generated images. We analyse user preferences over different clarification modes of text, image, and combination of both through crowdsourcing by taking into account image and text quality, clarity, and relevance. Our findings demonstrate that users generally prefer multi-modal clarification over uni-modal approaches. We explore the use of automated image generation techniques and compare the quality, relevance, and user preference of model-generated images with human-collected ones. The study reveals that text-to-image generation models, such as Stable Diffusion, can effectively generate multi-modal clarification questions. By investigating multi-modal clarification, this research establishes a foundation for future advancements in search systems.
Cite
@inproceedings{tavakoli2024understanding,
author = {Leila Tavakoli and Giovanni Castiglia and Federica Cal{\`o} and Yashar Deldjoo and Hamed Zamani and Johanne R. Trippas},
booktitle = {1st Workshop on Multimodal Search and Recommendations},
series = {CIKM MMSR '24},
title = {{Understanding Modality Preferences in Search Clarification}},
numpages = {9},
year = {2024}
}
Leif Azzopardi, Charles L. A. Clarke, Paul Kantor, Bhaskar Mitra, Johanne R. Trippas, and Zhaochun Ren. The Search Futures Workshop. In Proceedings of the European Conference on Information Retrieval (ECIR’24), ECIR ’24, pages 1–4, 2024.
Douglas W. Oard, Christopher Bearman, David Baker, Susannah Paletz, Johanne R. Trippas. Operational Disconnect in Mission Control. In [WS-14] Fearless Steps APOLLO Workshop, 2024.
2023
H. Zamani, J. R. Trippas, J. Dalton, and F. Radlinski. Conversational information seeking: An introduction to conversational search, recommendation, and question answering. Foundations and Trends in Information Retrieval, 2023.
[arXiv]
B. A. Martinez, R. Allmendinger, H. A. Khorshidi, T. Papamarkou, A. Feitas, J. R. Trippas, M. Zachariadis, N. Lord, and K. Benson. Applying artificial intelligence in fintech decision making to mitigate financial crime. In N. R. Vajjhala and K. D. Strang, editors, Cybersecurity for Decision Makers. Routlege/Taylor Francis/CRC Press, 2023.
P. Owoicho, J. Dalton, M. Aliannejadi, L. Azzopardi, J. R. Trippas, S. Vakulenko. TREC CAsT 2022: Going Beyond User Ask and System Retrieve with Initiative and Response Generation. Proceedings of the NIST Text Retrieval Conference (TREC 2022), TREC’22. pages 1–11, 2023.
S. Pathiyan Cherumanal, K. Ji, D. Hettiachchi, J. R. Trippas, Falk Scholer, Damiano Spina. RMIT_IR at the NTCIR-17 FairWeb-1 Task. Proceedings of 17th Conference on Evaluation of Information Access Technologies (NTCIR-17), 2023.
2022
L. Tavakoli, J. R. Trippas, H. Zamani, F. Scholer, and M. Sanderson. Mimics-duo: Offline online evaluation of search clarification. In Proceedings of the ACM Conference on Research and Development in Information Retrieval (SIGIR’22), pages 1–11, 2022.
[DOI]
J. Wei, B. Tag, J. R. Trippas, T. Dingler, and V. Kostakos. What could possibly go wrong? Understanding interaction errors with proactive smart speakers in the wild. In Proceedings of the ACM CHI Conference on Human Factors in Computing Systems (CHI’22), pages 1–20, 2022.
Y. Khaokaew, I. Holcombe-James, M. S. Rahaman, J. Liono, J. R. Trippas, D. Spina, P. Bailey, N. Belkin, P. N. Bennett, Y. Ren, M. Sanderson, F. Scholer, R. W. White, and F. D. Salim. Imagining future digital assistants at work: A study of task management needs. International Journal of Human-Computer Studies, 2022.
J. R. Trippas and D. Maxwell. First early career researchers’ roundtable for information access research. In Proceedings of the ACM Conference on Information Interaction and Retrieval (CHIIR’22), CHIIR ’22, pages 1–4, 2022.
[Roundtable website]
J. R. Trippas, D. Maxwell, A. Alqatan, M. Boon, C. Chavula, A. Crescenzi, L.-D. Ibanez, S. Meyer, A.-M. Ortloff, S. Palani, D. Patel, W. Thode, and Z. Xing. Report on the 1st Early Career Researchers’ Roundtable for Information Access Research (ECRs4IR 2022) at CHIIR 2022. SIGIR Forum, 56(1), 2022.
M. Aliannejadi and J. R. Trippas. Conversational information seeking: Theory and evaluation. In Proceedings of the ACM Conference on Information Interaction and Retrieval (CHIIR’22), CHIIR ’22, pages 1–2, 2022.
J. Dalton, S. Fischer, P. Owoicho, F. Radlinski, F. Rosetto, J. R. Trippas, and H. Zamani. Conversational information seeking: Theory and application. In Proceedings of the ACM Conference on Research and Development in Information Retrieval (SIGIR’22), pages 1–4, 2022.
2021
Y. Deldjoo, J. R. Trippas, and H. Zamani. Towards multi-modal conversational information seeking. In Proceedings of the ACM Conference on Research and Development in Information Retrieval (SIGIR’21), pages 1577–1587, 2021.
[DOI]
S. Gosper, J. R. Trippas, H. Richards, F. Allison, C. Sear, S. Khorasani, and F. Mattioli. Understanding the utility of digital flight assistants: A preliminary analysis. In Proceedings of the ACM International Conference on Conversational User Interfaces (CUI 2021), pages 1–5, 2021.
[DOI]
J. R. Trippas, D. Spina, M. Sanderson, and L. Cavedon. Accessing media via an audio-only communication channel: A log analysis. In Proceedings of the ACM International Conference on Conversational User Interfaces (CUI 2021), pages 1–6, 2021.
[DOI]
D. Spina, J. R. Trippas, P. Thomas, H. Joho, K. Byström, L. Clark, N. Craswell, M. Czerwinski, D. Elsweiler, A. Frummet, S. Ghosh, J. Kiesel, I. Lopatovska, D. McDuff, S. Meyer, A. Mourad, O. Paul, S. Pathiyan Cherumanal, D. Russell, and L. Sitbon. Future conversations. SIGIR Forum, 55(1), 2021.
[DOI]
J. R. Trippas and D. Maxwell. The PhD journey: Reaching out and lending a hand. In Proceedings of the ACM Conference on Information Interaction and Retrieval (CHIIR’21), CHIIR ’21, pages 345–346, 2021.
[DOI]
2020
J. R. Trippas, D. Spina, P. Thomas, H. Joho, M. Sanderson, and L. Cavedon. Towards a model for spoken conversational search. Information Processing & Management, 57(2):1–19, 2020.
[DOI] [Preprint]
J. Liono, M. S. Rahaman, F. D. Salim, Y. Ren, D. Spina, F. Scholer, J. R. Trippas, M. Sanderson, P. N. Bennett, and R. W. White. Intelligent task recognition: Towards enabling productivity assistance in daily life. In Proceedings of the 2020 International Conference on Multimedia Retrieval (ICMR), ICMR ’20, pages 472–478, 2020.
[DOI]
J. Mackenzie, R. Benham, M. Petri, J. R. Trippas, J. S. Culpepper, and A. Moffat. Cc-news-en: A large English news corpus. In Proceedings of the ACM International Conference on Information and Knowledge Management (CIKM2020), pages 3077–3084, 2020.
CC-News-EN dataset [DOI]
A. Vtyurina, C. Clarke, E. Law, J. R. Trippas, and H. Bota. A mixed-method analysis of text and audio search interfaces with varying task complexity. In Proceedings of the ACM Conference on International Conference on the Theory of Information Retrieval (ICTIR’20), ICTIR ’20, pages 61–68, 2020.
[DOI] [GitHub]
G. Buchanan, D. McKay, C. L. A. Clarke, L. Azzopardi, and J. R. Trippas. Made to measure: A workshop on human-centred metrics for information seeking. In Proceedings of the ACM Conference on Information Interaction and Retrieval (CHIIR’20), CHIIR ’20, pages 484–487, 2020.
J. R. Trippas, P. Thomas, D. Spina, and H. Joho. Third International Workshop on Conversational Approaches to Information Retrieval (CAIR’20). In Proceedings of the ACM Conference on Information Interaction and Retrieval (CHIIR’20), CHIIR ’20, pages 492–494, 2020.
[DOI]
2019
J. R. Trippas. Spoken Conversational Search: Audio-only Interactive Information Retrieval. RMIT University, 2019. (Thesis for Doctor of Philosopy (PhD), Science)
[SIGIR Forum doctoral abstract] [RMIT University Deputy Vice-Chancellor’s Higher Degree by Research Prize]
A. Chuklin, A. Severyn, J. R. Trippas, E. Alfonseca, H. Silen, and D. Spina. Using Audio Transformations to Improve Comprehension in Voice Question Answering. In Proceedings of the Conference and Labs of the Evaluation Forum (CLEF’19), pages 164–170, 2019.
[DOI]
CLEF2019-prosody GitHub
C. Qu, L. Yang, W. B. Croft, Y. Zhang, J. R. Trippas, and M. Qiu. User intent prediction in information-seeking conversations. In Proceedings of the ACM Conference on Information Interaction and Retrieval (CHIIR’19), pages 25–33, 2019.
[DOI]
J. R. Trippas, D. Spina, F. Scholer, A. Hassan Awadallah, P. Bailey, P. N. Bennett, R. W. White, J. Liono, Y. Ren, F. D. Salim, and M. Sanderson. Learning about work tasks to inform intelligent assistant design. In Proceedings of the ACM Conference on Information Interaction and Retrieval (CHIIR’19), pages 5–14, 2019.
[DOI]
J. Kim, J. R. Trippas, M. Sanderson, Z. Bao, and W. B. Croft. How do computer scientists use Google Scholar?: A survey of user interest in elements on SERPs and author profile pages. In Proceedings of the 8th Workshop on Bibliometric-enhanced Information Retrieval (BIR 2019). CEUR-WS, pages 64–75, 2019.
J. Liono, J. R. Trippas, D. Spina, M. S. Rahamad, Y. Ren, F. D. Salim, M. Sanderson, F. Scholer, and R. W. White. Building a Benchmark for Task Progress in Digital Assistants. In TI@WSDM10 WSDM Task Intelligence Workshop, pages 1–6, 2019.
J. R. Trippas and P. Thomas. Data sets for spoken conversational search. In Proceedings of the CHIIR 2019 Workshop on Barriers to Interactive IR Resources Re-use (BIIRRR 2019). CEUR-WS, pages 14–18, 2019.
2018
C. Qu, L. Yang, W. B. Croft, J. R. Trippas, Y. Zhang, and M. Qiu. Analyzing and characterizing user intent in information-seeking conversations. In Proceedings of the ACM Conference on Research and Development in Information Retrieval (SIGIR’18), pages 989– 992, 2018.
[DOI]
J. R. Trippas, D. Spina, L. Cavedon, H. Joho, and M. Sanderson. Informing the design of spoken conversational search. In Proceedings of the ACM Conference on Information Interaction and Retrieval (CHIIR’18), pages 32–41, 2018.
[DOI]
M. Aliannejadi, M. Hasanain, J. Mao, J. Singh, J. R. Trippas, H. Zamani, and L. Dietz. ACM SIGIR student liaison program. ACM SIGIR Forum, 51(3):42–45, 2018.
[DOI]
A. Chuklin, A. Severyn, J. R. Trippas, E. Alfonseca, H. Silen, and D. Spina. Prosody modifications for question-answering in voice-only settings. arXiv preprint arXiv:1806.03957, pages 1–5, 2018.
2017
D. Spina, J R. Trippas, L. Cavedon, and M. Sanderson. Extracting audio summaries to support effective spoken document search. Journal of the Association for Information Science and Technology, 68(9):2101–2115, 2017.
[DOI]
S. Shiga, H. Joho, R. Blanco, J. R. Trippas, and M. Sanderson. Modelling information needs in collaborative search conversations. In Proceedings of the ACM Conference on Research and Development in Information Retrieval (SIGIR’17), pages 715–724, 2017.
[DOI]
J. R. Trippas, Spina, L. Cavedon, and M. Sanderson. How do people interact in conversational speech-only search tasks: A preliminary analysis. In Proceedings of the ACM Conference on Information Interaction and Retrieval (CHIIR’17), pages 325–328, 2017.
[DOI] [Poster] [Dataset]
J. R. Trippas, D. Spina, L. Cavedon, and M. Sanderson. A conversational search transcription protocol and analysis. In CAIR’17 SIGIR 1st International Workshop on Conversational Approaches to Information Retrieval, pages 1–5, 2017.
J. R. Trippas, D. Spina, L. Cavedon, and M. Sanderson. Crowdsourcing user preferences and query judgments for speech-only search. In CAIR’17 SIGIR 1st International Workshop on Conversational Approaches to Information Retrieval, pages 1–3, 2017.
2016
J. R. Trippas. Spoken conversational search: Speech-only interactive information retrieval. In Proceedings of the ACM Conference on Information Interaction and Retrieval (CHIIR’16), pages 373–375, 2016.
[DOI]
2015
J. R. Trippas. Spoken conversational search: Information retrieval over a speech-only communication channel. In Proceedings of the ACM Conference on Research and Development in Information Retrieval (SIGIR’15), page 1067, 2015.
[DOI]
J. R. Trippas, D. Spina, M. Sanderson, and L. Cavedon. Towards understanding the impact of length in web search result summaries over a speech-only communication channel. In Proceedings of the ACM Conference on Research and Development in Information Retrieval (SIGIR’15), pages 991–994, 2015.
[DOI]
D. Spina, J. R. Trippas, L. Cavedon, and M. Sanderson. SpeakerLDA: Discovering topics in transcribed multi-speaker audio contents. In SLAM’15 Proceedings of ACM Multimedia 2015 Workshop on Speech, Language and Audio in Multimedia, pages 7–10, 2015.
[DOI]
Damiano Spina, Johanne R. Trippas, Lawrence Cavedon, and Mark Sanderson. SpeakerLDA: Discovering topics in transcribed multi-speaker audio contents. In SLAM’15 Proceedings of ACM Multimedia 2015 Workshop on Speech, Language and Audio in Multimedia, pages 7–10, 2015.
[PDF] [DOI] [Abstact] [Cite]
Abstract
Topic models such as Latent Dirichlet Allocation (LDA) have been extensively used for characterizing text collections according to the topics discussed in documents. Organizing documents according to topic can be applied to different information access tasks such as document clustering, content-based recommendation or summarization. Spoken documents such as podcasts typically involve more than one speaker (e.g., meetings, interviews, chat shows or news with reporters). This paper presents a work-in-progress based on a variation of LDA that includes in the model the different speakers participating in conversational audio transcripts. Intuitively, each speaker has her own background knowledge which generates different topic and word distributions. We believe that informing a topic model with speaker segmentation (e.g., using existing speaker diarization techniques) may enhance discovery of topics in multi-speaker audio content.
Cite
@inproceedings{spina2015speakerlda,
title={SpeakerLDA: Discovering topics in transcribed multi-speaker audio contents},
author={Spina, Damiano and Trippas, Johanne R and Cavedon, Lawrence and Sanderson, Mark},
booktitle={Proceedings of the Third Edition Workshop on Speech, Language \& Audio in Multimedia},
pages={7--10},
year={2015}
}
Johanne R. Trippas, Damiano Spina, Mark Sanderson, and Lawrence Cavedon. Results presentation methods for a spoken conversational search system. In NWSearch’15 First International Workshop on Novel Web Search Interfaces and Systems, pages 13–15, 2015.
[PDF] [DOI] [Abstract] [Cite]
Abstract
We propose research to investigate a new paradigm for Interactive Information Retrieval (IIR) where all input and output is mediated via speech. Our aim is to develop a new framework for effective and efficient IIR over a speech-only channel: a Spoken Conversational Search System (SCSS). This SCSS will provide an interactive conversational approach to determine user information needs, presenting results and enabling search reformulations. We have thus far investigated the format of results summaries for both audio and text, features such as summary length and summaries documents (noisy document or clean document) generated from (noisy) speech-recognition output from spoken document. In this paper we discuss future directions regarding a novel spoken interface targeted at search result presentation, query intent detection, and interaction patterns for audio search.
Cite
@inproceedings{trippas2015results,
author = {Trippas, Johanne R. and Spina, Damiano and Sanderson, Mark and Cavedon, Lawrence},
title = {Results Presentation Methods for a Spoken Conversational Search System},
year = {2015},
publisher = {Association for Computing Machinery},
address = {New York, NY, USA},
doi = {10.1145/2810355.2810356},
booktitle = {Proceedings of the First International Workshop on Novel Web Search Interfaces and Systems},
pages = {13--15},
numpages = {3},
location = {Melbourne, Australia}
}