Anhong Guo wearing a light blue shirt standing in front of a glass wall with drawings.

I design, develop, and deploy hybrid human-AI intelligent interactive systems to provide access to visual information in the real world. By combining the advantages of humans and AI, these systems can be nearly as robust and flexible as humans, and nearly as quick and low-cost as automated AI, enabling us to solve problems that are currently impossible with either alone.

I will start as an Assistant Professor at the University of Michigan CSE in January 2021!
research statement · teaching statement · diversity statement · curriculum vitae

Anhong Guo is an incoming Assistant Professor in Computer Science & Engineering at the University of Michigan starting January 2021. Anhong completed his Ph.D. in the Human-Computer Interaction Institute, Carnegie Mellon University, advised by Dr. Jeffrey Bigham. He is also a Snap Inc. Research Fellow, and a Swartz Innovation Fellow for Entrepreneurship. Anhong has published in many top academic conferences and journals on interface technologies, wearable computing, accessibility and computer vision, including two best paper nominees. Before CMU, he received his Master’s in HCI from Georgia Tech, and Bachelor's in Electronic Information Engineering from BUPT. He has also worked in the Ability and Intelligent User Experiences groups in Microsoft Research, the HCI group of Snap Research, the Accessibility Engineering team at Google, and the Mobile Innovation Center of SAP America.

News

Jul 6, 2020
Defending my dissertation on human-AI systems for visual information access
Jun 26, 2020
Three projects led by collaborators at CMU, MSR, and Apple accepted to ASSETS 2020
May-June, 2020
Serving on the program committees for UIST 2020 and ASSETS 2020
Apr 30, 2020
Joining University of Michigan CSE as an Assistant Professor in January 2021

Publications

Two example prototypes for making AR apps accessible. A: Foundational Accessibility. Screenshot of a virtual chair with a voice over target around it, a speech bubble shows the app announcing "Back of chair with blue cushion". B: Scanning. Screenshot of AR grid overlaid on a coffee table. Speech bubbles show the app announcing "Found a new horizontal surface" and "Scanned 2 surfaces totaling 2.3 square meters".
Jaylin Herskovitz, Jason Wu, Samuel White, Amy Pavel, Gabriel Reyes, Anhong Guo, and Jeffrey P. Bigham
ASSETS 2020
pdf · video (coming soon)
Screenshot of a tweet by @CDCgov from April 1, 2020 3:55pm: Actions to reduce spread of the virus, such as social distancing, are key to #FlattenTheCurve. 2 of 3 (original tweet link: https://twitter.com/CDCgov/status/1245439600472084486) The tweet contains an image of the common public health infographic about “flattening the curve”, but the tweet did not include alt text for the image. The image shows an example of a common flatten the curve info-graphic. A tall peak indicates the height of the pandemic if left unchecked, and a shorter spread out curve depicts the effects of social distancing efforts.
Cole Gleason, Stephanie Valencia, Lynn Kirabo, Jason Wu, Anhong Guo, Elizabeth J. Carter, Jeffrey P. Bigham, Cynthia L. Bennett, Amy Pavel
ASSETS 2020
The user is holding the phone in landscape mode with one hand, and aiming the camera towards a touchscreen coffee machine. The user’s other hand is wearing a fingercap exploring on the screen. The StateLens iOS app is providing audio guidance to the user.
Anhong Guo, Junhan Kong, Michael Rivera, Frank F. Xu, Jeffrey P. Bigham
UIST 2019
pdf · ACM DL · arXiv · full video · 30s preview · slides · talk · UIST talk
Two Blocks users are collaboratively creating a table in augmented reality.
Anhong Guo, Ilter Canberk, Hannah Murphy, Andrés Monroy-Hernández, Rajan Vaish
Ubicomp 2019
pdf · ACM DL · arXiv · video
On the left: a screenshot of Android App Drawer taken using X-Ray. On the right: a user holding the phone showing the same screenshot in X-Ray image viewer. Talk back cursor is visible.
Sujeath Pareddy, Anhong Guo, Jeffrey P. Bigham
Best Artifact AwardASSETS 2019
pdf · ACM DL · video · website
A word cloud composed of words including AI fairness, accessibility, artifical intelligence, inclusion, and bias.
Anhong Guo, Ece Kamar, Jennifer Wortman Vaughan, Hanna Wallach, Meredith Ringel Morris
ASSETS 2019 AI Fairness Workshop
Three images in the VizWiz-Priv dataset, including an image of a wall of photos containing faces, an image of a credit card, and an image of a pregnancy test. The private information regions in the images are highlighted and inpainted.
Danna Gurari, Qing Li, Chi Lin, Yinan Zhao, Anhong Guo, Abigale Stangl, Jeffrey P. Bigham
CVPR 2019
pdf · supplemental · CVF · website · poster
Interaction scenario of Minuet: after returning home, the user points at the Roomba and then the dirty area to ask Roomba to clean it up.
Runchang Kang, Anhong Guo, Gierad Laput, Yang Li, Xiang 'Anthony' Chen
SUI 2019
pdf · ACM DL · video
An example question sensor created in Zensors++ asking 'Is someone using a printer?' with a bounding box focusing on the printer area.
Anhong Guo, Anuraag Jain, Shomiron Ghose, Gierad Laput, Chris Harrison, Jeffrey P. Bigham
Ubicomp 2018
pdf · ACM DL · video · slides
A table with many boxes covered with white paper showing text such as glasses, butter, jam, etc. A user is holding and targeting his phone at one object, while touching the object with the other hand. This is showcasing the window cursor interaction technique that supports non-visual attention to items within a complex visual scene, in which the user moves the device itself to scan the scene and receives information about what is in the center of the image.
Anhong Guo, Saige McVea, Xu Wang, Patrick Clary, Ken Goldman, Yang Li, Yu Zhong, Jeffrey P. Bigham
ASSETS 2018
pdf · ACM DL · video · slides
Example printed overlays and legends generated by Facade. (a)-(d) demonstrate the different material combinations we tested in the design iterations (NinjaFlex with Braille, Flex+PLA Braille label, Flex+PLA Braille cover, and Flex+PLA embossed letter cover). Facade users can choose to print a legend for the abbreviations (e).
Anhong Guo, Jeffrey P. Bigham
IEEE Pervasive Computing 17(2), 2018, Maker Tech column
Distribution of the first six words for all questions in the VizWiz dataset. The innermost ring represents the first word and each subsequent ring represents a subsequent word. The arc size is proportional to the number of questions with that word/phrase.
Danna Gurari, Qing Li, Abigale J. Stangl, Anhong Guo, Chi Lin, Kristen Grauman, Jiebo Luo, Jeffrey P. Bigham
Spotlight PresentationCVPR 2018
pdf · arXiv · website · poster · video · tech review
A user accessing the microwave augmented with tactile overlays generated by Facade.
Anhong Guo, Jeeeun Kim, Xiang 'Anthony' Chen, Tom Yeh, Scott E. Hudson, Jennifer Mankoff, Jeffrey P. Bigham
CHI 2017
pdf · ACM DL · full video · 30s preview · talk
A user holding a 3D-printed cup holder augmented with a flexible ring generated using our flexible buffers technique.
Jeeeun Kim, Anhong Guo, Tom Yeh, Scott E. Hudson, Jennifer Mankoff
DIS 2017
pdf · ACM DL · video
The user is holding the phone in portrait mode with one hand, and aiming the camera towards an inaccessible microwave control panel. The user’s other hand is exploring on the panel. The VizLens iOS app is providing audio feedback and guidance to the user.
Anhong Guo, Xiang 'Anthony' Chen, Haoran Qi, Samuel White, Suman Ghosh, Chieko Asakawa, Jeffrey P. Bigham
UIST 2016
pdf · ACM DL · full video · 30s preview · talk
Two tilt-based interaction techniques for enabling no-touch, wrist-only interactions on smartwatches. Left: AnglePoint, which directly maps the position of a virtual pointer to the tilt angle of the smartwatch. Right: ObjectPoint, which objectifies the underlying virtual pointer as an object imbued with a physics model.
Anhong Guo, Tim Paek
Honorable MentionMobileHCI 2016
pdf · ACM DL · video
A user working on a document on a smartwatch using WearWrite, by leveraging a crowd to help trans­late their ideas into text.
Michael Nebeling, Alexandra To, Anhong Guo, Adrian A. de Freitas, Jaime Teevan, Steven P. Dow, Jeffrey P. Bigham
CHI 2016
pdf · ACM DL · full video · 30s preview · talk
A system architecture diagram of an order picking system augmented with weight checking error detection.
Xiaolong Wu, Malcolm Haynes, Anhong Guo, Thad Starner
ISWC 2016
pdf · ACM DL · video
Four BeyondTouch interaction techniques, including tapping on a phone in the pocket, tapping on the back of a phone while holding it with two hands, tapping and sliding on the back of the phone while holding it with one hand, as well as tapping and sliding next to the phone on the table to control the device.
Cheng Zhang, Anhong Guo, Dingtian Zhang, Yang Li, Caleb Southern, Rosa I. Arriaga, Gregory D. Abowd
TiiS 6(2), 2016
pdf · ACM DL
Pick-by-head-up display system using a Google Glass with a opaque display to show the pick order instructions.
Anhong Guo, Xiaolong Wu, Zhengyang Shen, Thad Starner, Hannes Baumann, Scott Gilliland
Computer 48(6), 2015
Four BeyondTouch interaction techniques, including tapping on a phone in the pocket, tapping on the back of a phone while holding it with two hands, tapping and sliding on the back of the phone while holding it with one hand, as well as tapping and sliding next to the phone on the table to control the device.
Cheng Zhang, Anhong Guo, Dingtian Zhang, Caleb Southern, Rosa I. Arriaga, Gregory D. Abowd
IUI 2015
pdf · ACM DL · video
Image of a set of order picking bins with LED displays and buttons.
Xiaolong Wu, Malcolm Haynes, Yixin Zhang, Ziyi Jiang, Zhengyang Shen, Anhong Guo, Thad Starner, Scott Gilliland
ISWC 2015
pdf · ACM DL
Image of the experimental setup, including 24 pick bins (on two shelving units with four rows and three columns each) and three order bins on the right. An example pick list is annotated with superimposed labels.
Anhong Guo, Shashank Raghu, Xuwen Xie, Saad Ismail, Xiaohui Luo, Joseph Simoneau, Scott Gilliland, Hannes Baumann, Caleb Southern, Thad Starner
Honorable MentionISWC 2014
pdf · ACM DL · video · talk