Number of co-authors:28
Number of publications with 3 favourite co-authors:Alexander G. Hauptmann:4Satanjeev Banerjee:3Nitendra Rajput:2
Alexander I. Rudnicky's 3 most productive colleagues in number of publications:Alexander G. Haupt..:43Dan R. Olsen Jr:28Nitendra Rajput:18
User error: replace user and press any key to continue.
-- Popular computer one-liner
Read the fascinating history of Wearable Computing, told by its father, Steve Mann
Read Steve's chapter !
Alexander I. Rudnicky
Has also published under the name of:
"A. I. Rudnicky"
Publications by Alexander I. Rudnicky (bibliography)
Sridharan, Seshadri, Chen, Yun-Nung, Chang, Kai-min and Rudnicky, Alexander I. (2012): NeuroDialog: an EEG-enabled spoken dialog interface. In: Proceedings of the 2012 International Conference on Multimodal Interfaces 2012. pp. 65-66.
Understanding user intent is a difficult problem in Dialog Systems, as they often need to make decisions under uncertainty. Using an inexpensive, consumer grade EEG sensor and a Wizard-of-Oz dialog system, we show that it is possible to detect system misunderstanding even before the user reacts vocally. We also present the design and implementation details of NeuroDialog, a proof-of-concept dialog system that uses an EEG based predictive model to detect system misrecognitions during live interaction.
© All rights reserved Sridharan et al. and/or ACM Press
Nanavati, Amit Anil, Rajput, Nitendra, Rudnicky, Alexander I. and Turunen, M. (2008): SiMPE: third workshop on speech in mobile and pervasive environments. In: Hofte, G. Henri ter, Mulder, Ingrid and Ruyter, Boris E. R. de (eds.) Proceedings of the 10th Conference on Human-Computer Interaction with Mobile Devices and Services - Mobile HCI 2008 September 2-5, 2008, Amsterdam, the Netherlands. pp. 569-570.
Banerjee, Satanjeev and Rudnicky, Alexander I. (2007): Segmenting meetings into agenda items by extracting implicit supervision from human note-taking. In: Proceedings of the 2007 International Conference on Intelligent User Interfaces 2007. pp. 151-159.
Splitting a meeting into segments such that each segment contains discussions on exactly one agenda item is useful for tasks such as retrieval and summarization of agenda item discussions. However, accurate topic segmentation of meetings is a difficult task. In this paper, we investigate the idea of acquiring implicit supervision from human meeting participants to solve the segmentation problem. Specifically we have implemented and tested a note taking interface that gives value to users by helping them organize and retrieve their notes easily, but that also extracts a segmentation of the meeting based on note taking behavior. We show that the segmentation so obtained achieves a Pk value of 0.212 which improves upon an unsupervised baseline by 45% relative, and compares favorably with a current state-of-the-art algorithm. Most importantly, we achieve this performance without any features or algorithms in the classic sense.
© All rights reserved Banerjee and Rudnicky and/or ACM Press
Nanavati, A. A., Rajput, N., Rudnicky, Alexander I. and Sicconi, R. (2006): SiMPE: speech in mobile and pervasive environments. In: Proceedings of 8th conference on Human-computer interaction with mobile devices and services 2006. pp. 293-294.
Traditionally, voice-based applications have been accessed using unintelligent telephone devices through Voice Browsers that reside on the server. The proliferation of pervasive devices and the increase in their processing capabilities, client-side speech processing is emerging as a viable alternative. This workshop will explore the various possibilities and issues that arise while enabling speech processing on resource-constrained, possibly mobile devices. The workshop will highlight the many open areas that require research attention, identify key problems that need to be addressed, and also discuss a few approaches for solving some of them - to build the next generation of conversational systems.
© All rights reserved Nanavati et al. and/or ACM Press
Nanavati, Amit Anil, Rajput, Nitendra, Rudnicky, Alexander I. and Sicconi, Roberto (2006): SiMPE: speech in mobile and pervasive environments. In: Nieminen, Marko and Röykkee, Mika (eds.) Proceedings of the 8th Conference on Human-Computer Interaction with Mobile Devices and Services - Mobile HCI 2006 September 12-15, 2006, Helsinki, Finland. pp. 293-294.
Banerjee, S., Rose, C. and Rudnicky, Alexander I. (2005): The Necessity of a Meeting Recording and Playback System, and the Benefit of Topic-Level Annotations to Meeting Browsing. In: Proceedings of IFIP INTERACT05: Human-Computer Interaction 2005. pp. 643-656.
Much work in the area of Computer Supported Cooperative Work (CSCW) has targeted the problem of supporting meetings between collaborators who are non-collocated, enabling meetings to transcend boundaries of space. In this paper, we explore the beginnings of a proposed solution for allowing meetings to transcend time as well. The need for such a solution is motivated by a user survey in which busy professionals are questioned about meetings they have either missed or forgotten the important details about after the fact. Our proposed solution allows these professionals to transcend time in a sense by revisiting a recorded meeting that has been structured for quick retrieval of sought information. Such a solution supports complete recovery of prior discussions, allowing needed information to be retrieved quickly, and thus potentially facilitating the effective continuation of discussions from the past. We evaluate the proposed solution with a formal user study in which we measure the impact of the proposed structural annotations on retrieval of information. The results of the study show that participants took significantly less time to retrieve the answers when they had access to discourse structure based annotation than in a control condition in which they had access only to unannotated video recordings (p < 0.01, effect size 0.94 standard deviations).
© All rights reserved Banerjee et al. and/or Springer Verlag
Rybski, Paul E., Banerjee, Satanjeev, Torre, Fernando De la, Vallespí, Carlos, Rudnicky, Alexander I. and Veloso, Manuela M. (2004): Segmentation and classification of meetings using multiple information streams. In: Sharma, Rajeev, Darrell, Trevor, Harper, Mary P., Lazzari, Gianni and Turk, Matthew (eds.) Proceedings of the 6th International Conference on Multimodal Interfaces - ICMI 2004 October 13-15, 2004, State College, PA, USA. pp. 335-336.
Rybski, Paul E., Banerjee, Satanjeev, Torre, Fernando De la, Vallespí, Carlos, Rudnicky, Alexander I. and Veloso, Manuela (2004): Segmentation and classification of meetings using multiple information streams. In: Proceedings of the 2004 International Conference on Multimodal Interfaces 2004. pp. 335-336.
We present a meeting recorder infrastructure used to record and annotate events that occur in meetings. Multiple data streams are recorded and analyzed in order to infer a higher-level state of the group's activities. We describe the hardware and software systems used to capture people's activities as well as the methods used to characterize them.
© All rights reserved Rybski et al. and/or their publisher
Rosenfeld, Ronald, Olsen Jr, Dan R. and Rudnicky, Alexander I. (2001): Universal speech interfaces. In Interactions, 8 (6) pp. 34-44.
Hauptmann, Alexander G., Witbrock, Michael J., Rudnicky, Alexander I. and Reed, Stephen (1995): Speech for Multimedia Information Retrieval. In: Robertson, George G. (ed.) Proceedings of the 8th annual ACM symposium on User interface and software technology November 15 - 17, 1995, Pittsburgh, Pennsylvania, United States. pp. 79-80.
We describe the Informedia News-on-Demand system. News-on-Demand is an innovative example of indexing and searching broadcast video and audio material by text content. The fully-automatic system monitors TV news and allows selective retrieval to news items based on spoken queries. The user then plays the appropriate video "paragraph". The system runs on a Pentium PC using MPEG-I video compression and the Sphinx-II continuous speech recognition system .
© All rights reserved Hauptmann et al. and/or ACM Press
Rudnicky, Alexander I., Hauptmann, Alexander G. and Lee, Kai-Fu (1994): Survey of Current Speech Technology. In Communications of the ACM, 37 (3) pp. 52-57.
Teal, Steven L. and Rudnicky, Alexander I. (1992): A Performance Model of System Delay and User Strategy Selection. In: Bauersfeld, Penny, Bennett, John and Lynch, Gene (eds.) Proceedings of the ACM CHI 92 Human Factors in Computing Systems Conference June 3-7, 1992, Monterey, California. pp. 295-305.
This study lays the ground work for a predictive, zero-parameter engineering model that characterizes the relationship between system delay and user performance. This study specifically investigates how system delays affects a user's selection of task strategy. Strategy selection is hypothesized to be based on a cost function combining two factors: (1) the effort required to synchronize input with system availability and (2) the accuracy level afforded. Results indicate that users, seeking to minimize effort and maximize accuracy, choose among three strategies -- automatic performance, pacing, and monitoring. These findings provide a systematic account of the influence of system delay on user performance, based on adaptive strategy choice drive by cost.
© All rights reserved Teal and Rudnicky and/or ACM Press
Rudnicky, Alexander I. and Hauptmann, Alexander G. (1991): Models for Evaluating Interaction Protocols in Speech Recognition. In: Robertson, Scott P., Olson, Gary M. and Olson, Judith S. (eds.) Proceedings of the ACM CHI 91 Human Factors in Computing Systems Conference April 28 - June 5, 1991, New Orleans, Louisiana. pp. 285-291.
Recognition errors complicate the assessment of speech systems. This paper presents a new approach to modeling spoken language interaction protocols, based on finite Markov chains. An interaction protocol, prescribed by the interface design, defines a set of primitive transaction steps and the order of their execution. The efficiency of an interface depends on the interaction protocol as well as the cost of each different transaction step. Markov chains provide a simple and computationally efficient method for modeling errorful systems. They allow for detailed comparisons between different interaction protocols and between different modalities. The method is illustrated by application to example protocols.
© All rights reserved Rudnicky and Hauptmann and/or ACM Press
Lunati, Jean-Michel and Rudnicky, Alexander I. (1991): Spoken Language Interfaces: The OM System. In: Robertson, Scott P., Olson, Gary M. and Olson, Judith S. (eds.) Proceedings of the ACM CHI 91 Human Factors in Computing Systems Conference April 28 - June 5, 1991, New Orleans, Louisiana. pp. 453-454.
Teal, Steven L. and Rudnicky, Alexander I. (1991): Changes in User Task Strategy Due to System Response Delay. In ACM SIGCHI Bulletin, 23 (4) pp. 47-48.
Rudnicky, Alexander I., Sakamoto, Michelle and Polifroni, Joseph H. (1990): Spoken Language Interaction in a Spreadsheet Task. In: Diaper, Dan, Gilmore, David J., Cockton, Gilbert and Shackel, Brian (eds.) INTERACT 90 - 3rd IFIP International Conference on Human-Computer Interaction August 27-31, 1990, Cambridge, UK. pp. 767-772.
To study the spoken language interface in the context of a complex problem-solving task, we had a group of users perform a spreadsheet task, alternating voice and keyboard input. A total of 40 tasks were performed by each participant, the first thirty in a group (over several days), the remaining ones a month later. The voice spreadsheet program used in this study was extensively instrumented to provide detailed information about the components of the interaction. These data, as well as analysis of the participants' utterances and recognizer output, provide a fairly detailed picture of spoken language interaction.
© All rights reserved Rudnicky et al. and/or North-Holland
Hauptmann, Alexander G. and Rudnicky, Alexander I. (1988): Talking to Computers: An Empirical Investigation. In International Journal of Man-Machine Studies, 28 (6) pp. 583-604.
This paper describes an empirical study of man-computer speech interaction. The goals of the experiment were to find out how people would communicate with a real-time, speaker-independent continuous speech understanding system. The experimental design compared three communication modes: natural language typing, speaking directly to a computer and speaking to a computer through a human interpreter. The results show that speech to a computer is not as ill-formed as one would expect. People speaking to a computer are more disciplined than when speaking to each other. There are significant differences in the usage of spoken language compared to typed language, and several phenomena which are unique to spoken or typed input respectively. Usefulness for work in speech understanding systems for the future is considered.
© All rights reserved Hauptmann and Rudnicky and/or Academic Press
Show this list on your homepage
Join the technology elite and advance:
Changes to this page (author)10 Nov 2012: Added20 Apr 2011: Added
16 Feb 2010: Modified
18 Aug 2009: Added
30 May 2009: Added
29 May 2009: Added
29 May 2009: Added
24 Jul 2007: Added
24 Jul 2007: Added
24 Jul 2007: Added
23 Jun 2007: Added
28 Apr 2003: Added
Page maintainer: The Editorial Team