Number of co-authors:15
Number of publications with 3 favourite co-authors:David R. Karger:2Björn Hartmann:1Robert C. Miller:1
Katrina Panovich's 3 most productive colleagues in number of publications:Mark S. Ackerman:67Robert C. Miller:42Meredith Ringel Mo..:38
For a list of all the ways technology has failed to improve the quality of life, please press three.
-- Alice Kahn
Read the fascinating history of Wearable Computing, told by its father, Steve Mann
Read Steve's chapter !
Publications by Katrina Panovich (bibliography)
Panovich, Katrina, Miller, Rob and Karger, David (2012): Tie strength in question & answer on social network sites. In: Proceedings of ACM CSCW12 Conference on Computer-Supported Cooperative Work 2012. pp. 1057-1066.
Asking friends, colleagues, or other trusted people to help answer a question or find information is a familiar and tried-and-true concept. Widespread use of online social networks has made social information seeking easier, and has provided researchers with opportunities to better observe this process. In this paper, we relate question answering to tie strength, a metric drawn from sociology describing how close a friendship is. We present a study evaluating the role of tie strength in question answers. We used previous research on tie strength in social media to generate tie strength information between participants and their answering friends, and asked them for feedback about the value of answers across several dimensions. While sociological studies have indicated that weak ties are able to provide better information, our findings are significant in that weak ties do not have this effect, and stronger ties (close friends) provide a subtle increase in information that contributes more to participants' overall knowledge, and is less likely to have been seen before.
© All rights reserved Panovich et al. and/or ACM Press
Morris, Meredith Ringel, Teevan, Jaime and Panovich, Katrina (2010): What do people ask their social networks, and why?: a survey study of status message q&a behavior. In: Proceedings of ACM CHI 2010 Conference on Human Factors in Computing Systems 2010. pp. 1739-1748.
People often turn to their friends, families, and colleagues when they have questions. The recent, rapid rise of online social networking tools has made doing this on a large scale easy and efficient. In this paper we explore the phenomenon of using social network status messages to ask questions. We conducted a survey of 624 people, asking them to share the questions they have asked and answered of their online social networks. We present detailed data on the frequency of this type of question asking, the types of questions asked, and respondents' motivations for asking their social networks rather than using more traditional search tools like Web search engines. We report on the perceived speed and quality of the answers received, as well as what motivates people to respond to questions seen in their friends' status messages. We then discuss the implications of our findings for the design of next-generation search tools.
© All rights reserved Morris et al. and/or their publisher
Bernstein, Michael S., Little, Greg, Miller, Robert C., Hartmann, Björn, Ackerman, Mark S., Karger, David R., Crowell, David and Panovich, Katrina (2010): Soylent: a word processor with a crowd inside. In: Proceedings of the 2010 ACM Symposium on User Interface Software and Technology 2010. pp. 313-322.
This paper introduces architectural and interaction patterns for integrating crowdsourced human contributions directly into user interfaces. We focus on writing and editing, complex endeavors that span many levels of conceptual and pragmatic activity. Authoring tools offer help with pragmatics, but for higher-level help, writers commonly turn to other people. We thus present Soylent, a word processing interface that enables writers to call on Mechanical Turk workers to shorten, proofread, and otherwise edit parts of their documents on demand. To improve worker quality, we introduce the Find-Fix-Verify crowd programming pattern, which splits tasks into a series of generation and review stages. Evaluation studies demonstrate the feasibility of crowdsourced editing and investigate questions of reliability, cost, wait time, and work time for edits.
© All rights reserved Bernstein et al. and/or their publisher
Kleek, Max G. Van, Bernstein, Michael, Panovich, Katrina, Vargas, Gregory G., Karger, David R. and Schraefel, MC (2009): Note to self: examining personal information keeping in a lightweight note-taking tool. In: Proceedings of ACM CHI 2009 Conference on Human Factors in Computing Systems 2009. pp. 1477-1480.
This paper describes a longitudinal field experiment in personal note-taking that examines how people capture and use information in short textual notes. Study participants used our tool, a simple browser-based textual note-taking utility, to capture personal information over the course of ten days. We examined the information they kept in notes using the tool, how this information was expressed, and aspects of note creation, editing, deletion, and search. We found that notes were recorded extremely quickly and tersely, combined information of multiple types, and were rarely revised or deleted. The results of the study demonstrate the need for a tool such as ours to support the rapid capture and retrieval of short notes-to-self, and afford insights into how users' actual note-keeping tendencies could be used to better support their needs in future PIM tools.
© All rights reserved Kleek et al. and/or ACM Press
Show this list on your homepage
Join the technology elite and advance:
Changes to this page (author)03 Apr 2012: Added03 Nov 2010: Added
02 Nov 2010: Added
25 Feb 2010: Modified
09 May 2009: Added
Page maintainer: The Editorial Team