AI2 at ACM UIST 2023

New Intelligent Reading Interfaces Research and The Semantic Reader Open Research Platform

Joseph Chee Chang
AI2 Blog

--

This year at the 36th Annual ACM UIST Conference, researchers from AI2 will present papers on two new AI-powered intelligent reading interfaces for scholarly documents — Papeo and Synergi. Coinciding with the conference, we are also releasing to the broader research community The Semantic Reader Open Research Platform, a set of open-source libraries that powered the development of Papeo, Synergi, and a total of more than 10 reading interfaces research we have published in recent years.

An image with a large title that said “UIST 2023” in the foreground, and in the background you can see the golden gate bridge, the sky is clear and the sun is just setting a little behind the bridge, giving the plants in front of the bridge a warm glow. Light fog over the water.
The ACM Symposium on User Interface Software and Technology (UIST) is the premier forum for innovations in Human-Computer Interfaces. Sponsored by ACM special interest groups on computer-human interaction (SIGCHI) and computer graphics (SIGGRAPH), UIST brings together people from diverse areas including Human-Centered AI, graphical & web user interfaces, Computer-Supported Cooperative Work, and more.

Highlighted UIST Papers from AI2 Researchers

(*AI2 affiliation, +Work completed during a research internship at AI2)

Papeos: Augmenting Research Papers with Talk Videos

Tae Soo Kim+, Matt Latzke,* Jonathan Bragg,* Amy X. Zhang,* Joseph Chee Chang.*

Read more about Papeos, including interactive demos and talk videos.

Research consumption has largely involved reading academic papers, which are often static, dense, and formally written. Alternatively, pre-recorded conference presentation videos have recently become more widely available. In contrast to research papers, talk videos offer a more dynamic, concise, and colloquial style for conveying research. Despite their complementary benefits, talk videos are potentially under utilized after the conferences.

A research paper on the left, with vertical color bars in the margin next to some of the passages. A dotted line link each of the vertical bars to a corresponded video clip to the right of the paper. The video clips are also color coded to showed which passage they were relevant to. All the way to the right, it showed that the video clips are all part of a longer video that is the author’s full conference talk presentation video.
Papeos augment academic papers by linking relevant passages and segments of authors’ talk videos. Video segments are presented as margin notes that are localized and color-coded next to relevant passages. Users can fluidly switch between consuming the dense and detailed paper text, and the typically more concise and colloquial talk video — -providing a new scholarly reading experience.

Our research delves into the potential advantages of blending academic papers with talk videos to create a more engaging and dynamic research consumption experience. Through formative and co-design studies, we designed Papeos, a novel reading and intelligent authoring interface that allow authors to augment their papers by segmenting and localizing talk videos alongside relevant paper passages. Papeos enables readers to visually skim a paper through clip thumbnails, and fluidly switch between consuming dense text in the paper or visual summaries in the video. Our comparative lab study showed that Papeos reduced mental load, scaffolded navigation, and facilitated more comprehensive reading of papers.

Figure 4 shows diagrams of three main interactions in Papeo which are shown left-to-right. Figure 4A (Left) shows, at the top, a screenshot of the Papeo Reader where the cursor is hovering over a highlight bar with a video note to the side. At the bottom, a screenshot that illustrates the cursor moving down the same highlight bar and the video note now shows a different thumbnail. Figure 4B (Middle) shows an activated video note which displays a segmented timeline at the top where the user is ho
Illustration of features supported by the Papeo reader:

Synergi: A Mixed-Initiative System for Scholarly Synthesis and Sensemaking

Hyeonsu B. Kang, Sherry Tongshuang Wu, Joseph Chee Chang,* Aniket Kittur.

Read more about Synergi, including demo and talk videos.

The four main stages of Synergi which are 1) Clipping via Highlighting, 2) Local Citation Graph Construction and Loopy Belief Propagation over it, 3) Clip Extraction, Clustering, and Recursive Summarization using Chat-GPT4, and 4) Importing to and Modifying in the Interactive Outline Editor.
[1] A scholar highlights a patch of text and references that describes an interesting research problem. [2] Loopy Belief Propagation is performed on the citation graph to retrieves important papers and quotes within 2-hops relevant to the highlighted text. [3] GPT4 is used to iteratively summarize the retrieved text into hierarchical topics that [4] the scholar can explore and intergrade into their outlines.

Efficiently reviewing and weaving together scholarly literature is crucial for advancing scientific knowledge. However, the ever-expanding volume of publications and the weight of accumulated knowledge pose a formidable challenge to the synthesis of research threads in the literature. While significant research has been devoted to helping scholars interact with individual papers, building research threads scattered across multiple papers remains a challenge. One core challenge here is that top-down synthesis methods and large language models have their limitations in terms of personalization and adaptability, whereas bottom-up synthesis is laborious and time-consuming.

In response to these challenges, we introduce an innovative approach involving mixed-initiative workflows. We introduce Synergi, a novel computational pipeline and interface that allows users to select descriptions of cited papers in a reading interface as seeds, and exploit both the larger citation graphs and LLMs to expand and summarize an additional set of relevant research threads and papers for users to explore. Synergi expowers scholars to start with an entire threads-and-subthreads structure generated from papers relevant to their interests, and to iterate and customize on it as they wish. Our evaluation showed that Synergi helps scholars efficiently make sense of relevant threads, broaden their perspectives, and increase their curiosity. We discuss future design implications for thread-based, mixed-initiative scholarly synthesis support tools.

The Semantic Reader Open Research Platform

A figure with the title “The Semantic Reader Open Research Platform” and a brief description that mostly overlaps with the paragraph below. At the bottom there is a list of institutions: AI2, UW, UC Berkeley, Upenn, MIT, UIUC, CMU, and UMinnesota.

More broadly, Papeos and Synergi are a part of the Semantic Reader Project, a collaborative effort of NLP + HCI researchers from non-profit, industry, and academic institutions to create AI-powered interactive and intelligent reading interfaces for scholarly papers. Our research led to the creation of Semantic Reader, an application that incorporates novel research features as they mature, and is used by tens of thousands of scholars each week on the Semantic Scholar website.

Finally, coinciding with UIST 2023, we are also releasing The Semantic Reader Open Research Platform. With the platform, we open source NLP toolkits and React UI libraries that we have developed to power more than a dozen intelligent reading interface prototypes. More specifically, we are releasing PaperMage (to appear at EMNLP 2023), a library for processing and analyzing scholarly PDFs, and PaperCraft, a React UI component for building augmented and interactive reading interfaces. We also included documentations on how to use these resources to build research prototypes. For example, on the website you can find a detailed tutorial on building Paper Plain, an intelligent reading interface with in-situ term definitions, passage summaries, and suggested key questions to guide laypeople when reading medical papers. Paper Plain was published in ACM TOCHI 2023, and will be presented at the upcoming ACM SIGCHI 2024.

We hope by releasing these resources, they can enable the broader research community to explore exciting challenges around novel research support tools. Join us in designing the future of scholarly reading interfaces with our open source libraries!

Read more about the Semantic Reader Open Research Platform, including interactive demos, source codes, and tutorials on how to build intelligent scholarly reading interfaces here.

References

PaperMage: A Unified Toolkit for Processing, Representing, and Manipulating Visually-Rich Scientific Documents. Kyle Lo, Zejiang Shen, Benjamin Newman, Joseph Chee Chang, Russell Authur, Erin Bransom, Stefan Candra, Yoganand Chandrasekhar, Regan Huff, Bailey Kuehl, Amanpreet Singh, Chris Wilhelm, Angele Zamarron, Marti A. Hearst, Daniel S. Weld, Doug Downey, Luca Soldaini. (To appear) Conference on Empirical Methods in Natural Language Processing: Demos. 2023.

Follow @allen_ai and @semanticscholar on Twitter/X, and subscribe to the AI2 Newsletter to stay current on news and research coming out of AI2.

--

--

😉💻🔄 Research Scientist @ AI2/Semantic Scholar | prev @ Carnegie Mellon