Co-chairing AutomotiveUI 2010

On November 11 and 12 I was at the AutomotiveUI 2010 conference serving as program co-chair with Susanne Boll. The conference was hosted by Anind Dey at CMU and co-chaired by Albrecht Schmidt.

The conference was successful and really fun. I could go on about all the great papers and posters (including two posters from our group at UNH [1,2]) but in this post I’ll only mention two: John Krumm’s keynote talk and, selfishly, my own talk (this is my blog after all). John gave an overview of his work with data from GPS sensors. He discussed work on prediction of where people will go, his experiences with location privacy and with creating road maps. Given that John is, according to his own website, the “all seeing, all knowing, master of time, space, and dimension,” this was indeed a very informative talk 😉 OK in all seriousness, the talk was excellent. I find John’s work on prediction of people’s destination and selected route the most interesting. One really interesting effect of having accurate predictions, and people sharing such data in the cloud, would be on routing algorithms hosted in the cloud. If such an algorithm could know where all of us are going at any instant of time, it could propose routes that overall allow efficient use of roads, reduced pollution, etc.

My talk focused on collaborative work with Alex Shyrokov and Peter Heeman on multi-threaded dialogues. Specifically, I talked about designing spoken tasks for human-human dialogue experiments for Alex’s PhD work [3]. Alex wanted to observe how pairs of subjects switch between two dialogue threads, while one of the subjects is also engaged in operating a simulated vehicle. Our hypothesis is that observed human-human dialogue behaviors can be used as the starting point for designing computer dialogue behaviors for in-car spoken dialogue systems. One of the suggestions we put forth in the paper is that the tasks for human-human experiments should be engaging. These are the types of tasks that will result in interesting dialogue behaviors and can thus teach us something about how humans manage multi-threaded dialogues.

Next year the conference moves back to Europe. The host will be Manfred Tscheligi in Salzburg, Austria. Judging by the number of submissions this year and the quality of the conference, we can look forward to many interesting papers next year, both from industry and from academia. Also, the location will be excellent – just think Mozart, Sound of Music (see what Rick Steves has to say), and world-renowned Christmas markets!


[1] Zeljko Medenica, Andrew L. Kun, Tim Paek, Oskar Palinko, “Comparing Augmented Reality and Street View Navigation,” AutomotiveUI 2010 Adjunct Proceedings

[2] Oskar Palinko, Sahil Goyal, Andrew L. Kun, “A Pilot Study of the Influence of Illumination and Cognitive Load on Pupil Diameter in a Driving Simulator,” AutomotiveUI 2010 Adjunct Proceedings

[3] Andrew L. Kun, Alexander Shyrokov, Peter A. Heeman, “Spoken Tasks for Human-Human Experiments: Towards In-Car Speech User Interfaces for Multi-Threaded Dialogue,” AutomotiveUI 2010

Report on ECE Graduate Seminar

During the 2009-2010 academic year I taught a new version of the UNH ECE Graduate Seminar (ECE 900), a course I first introduced in the fall of 2002. The primary aim of the previous version of the course was to expose our graduate students to research and development conducted at other institutions. Thus, the course consisted of eight invited talks per semester, given by engineers, scientist and other professionals, and covering a range of topics of interest to ECE students. 

Starting with the 2009-2010 academic year the primary aim of the course has become to introduce graduate students to the general tools of scientific research. I championed this new aim for the course and I’m grateful that my faculty colleagues gave me an opportunity to share my excitement about scientific research with our graduate students. 

The course had three main aspects: 

  • Lectures on performing scientific research. My lectures introduced students to the steps of scientific research, from formulating problems, to proposing hypotheses and conducting experiments.
  • Research talks. Each student attended at least 15 talks. Most of these were held at UNH and the speakers were exceptional. At the same time, I encouraged the students to recommend talks that we can attend at other institutions. The result: trips to MIT, BU and WPI.
  • Research proposal. At the end of the two semester sequence each student submitted a short research proposal and gave a presentation on the same. The proposals were developed over the two semesters, with students working individually and in groups. I provided feedback throughout the year on different segments of the proposal. 

In an informal survey at the end of the academic year most students indicated that they liked the new version of the ECE Graduate Seminar and that they thought it was useful. All of the students thought that learning about the tools of science is useful and the majority also indicated that their technical writing skills improved due to this course. These responses are certainly encouraging. 

I will be teaching ECE 900 again during the 2010-2011 academic year. Based on my experiences reported here, as well as those with my Fundamentals of Ubicomp course, I plan to implement two changes: 

  • Accelerate proposal development. I will move up the due date for the final research proposal to sometime early in the second semester. The accelerated schedule should help build excitement for learning about science. It will also give us time at the end of the year to discuss how other researchers approach scientific work. Finally, it will help with student participation in the course, which is the subject of the second change I intend to implement.
  • Increase student participation. While I encouraged student participation throughout the semester, the results were not always stellar. By accelerating the proposal development process I hope to provide students with discussion topics that they feel comfortable talking about. I also intend to ask students to hold multiple formal presentations in class. One assignment that students can expect next semester: create a 15 minute presentation about a research topic of your choice, based on a research video posted online.

Talk at SpeechTEK 2010

On Tuesday (August 3, 2010) I attended SpeechTEK 2010. I had a chance to see several really interesting talks including the lunch keynote by Zig Serafin, General Manager, Speech at Microsoft. He and two associates discussed, among other topics, the upcoming release of the Windows 7 phone and of the Kinect for Xbox 360 (formerly Project Natal). We also saw successful live demonstrations of both of these technologies.

One of Zig’s associates to take the stage was Larry Heck, Chief Scientist, Speech at Microsoft. Larry believes that there are three areas of research and development that will combine to make speech a part of everyday interactions with computers. First, the advent of ubiquitous computing and the need for natural user interfaces (NUIs) means that we cannot keep relying on GUIs and keyboards for many of our computing needs. Second, cloud computing makes it possible to gather rich data to train speech systems. Finally, with advances in speech technology we can expect to see search move beyond typing keywords (which is what we do today sitting at our PCs) to conversational queries (which is what people are starting to do on mobile phones).

I attended four other talks with topics relevant to my research. Brigitte Richardson discussed her work on Ford’s Sync. It’s exciting to hear that Ford is coming out with an SDK that will allow integrating devices with Sync. This appears to be a similar approach to ours at Project54 – we also provide an SDK which can be used to write software for the Project54 system [1]. Eduardo Olvera of Nuance discussed the differences and similarities between designing interfaces for speech interaction and those for interaction on a small form factor screen. Karen Kaushansky of TellMe discussed similar issues focusing on customer care. Finally, Kathy Lee, also of TellMe, discussed her work on a diary study exploring when people are willing to talk to their phones. This work reminded me of an experiment in which Ronkainen et al. asked participants to rate the social acceptability of mobile phone usage scenarios they viewed in video clips [2].

I also had a chance to give a talk reviewing some of the results of my collaboration with Tim Paek of Microsoft Research. Specifically, I discussed the effects of speech recognition accuracy and PTT button usage on driving performance [3] and the use of voice-only instructions for personal navigation devices [4]. The talk was very well received by the audience of over 25, with many follow-up questions. Tim also gave this talk earlier this year at Mobile Voice 2010.

For pictures from SpeechTEK 2010 visit my Flickr page.


[1] Andrew L. Kun, W. Thomas Miller, III, Albert Pelhe and Richard L. Lynch, “A software architecture supporting in-car speech interaction,” IEEE Intelligent Vehicles Symposium 2004

[2] Sami Ronkainen, Jonna Häkkilä, Saana Kaleva, Ashley Colley, Jukka Linjama, “Tap Input as an Embedded Interaction Method for Mobile Devices,” TEI 2007

[3] Andrew L. Kun, Tim Paek, Zeljko Medenica, “The Effect of Speech Interface Accuracy on Driving Performance,” Interspeech 2007

[4] Andrew L. Kun, Tim Paek, Zeljko Medenica, Nemanja Memarovic, Oskar Palinko, “Glancing at Personal Navigation Devices Can Affect Driving: Experimental Results and Design Implications,” Automotive UI 2009

Report on Fundamentals of Ubicomp course

During the spring 2010 semester I taught a new course entitled Ubiquitous Computing Fundamentals. The term ubiquitous computing refers to the model of computing in which computers are embedded in everyday objects and become part of everyday activities. As the name implies, this course was designed as an introduction to this exciting field of study.

In this course I used the excellent new ubicomp textbook [1] edited by John Krumm. I highly recommend this book to anyone starting out in the field of ubicomp. Specifically, I like two aspects of the book. First, the team of contributors assembled by John provides a comprehensive introduction to the myriad topics that make up the ubicomp field. The fact that ubicomp is an interdisciplinary field is exciting, but getting an overview of the field may seem like a daunting task. The textbook provides this overview. Second, paraphrasing Aaron Quigley‘s assessment of his chapter [2], the book provides “an entry point” to the world of conducting research in general, and ubicomp research in particular. The contributors discuss the tools used in various aspects of ubicomp research, from prototyping, to user studies, to data processing. The individual chapters help the reader formulate research questions and steps, and provide valuable tips on how to report on results. 

The course covered three topic areas:

  • History of ubicomp. The semester started with Weiser’s seminal paper [3] and with a textbook chapter introducing ubicomp by Roy Want, one of Weiser’s collaborators at Xerox PARC.
  • Building ubicomp systems. We discussed various aspects of creating ubicomp systems, from writing always-on software, to privacy, to conducting laboratory and field experiments.
  • The user experience. As this is my research focus, we spent a considerable amount of time discussion user interactions with ubicomp systems, from speech interactions, to multi-touch tables, to tangible user interfaces.

I found that an excellent way to discuss ubicomp topics is to take advantage of research videos posted online. We viewed many such videos and this led to productive discussions. We also benefited from excellent talks by Marko PopovicBret Harsham and Albrecht Schmidt.

I felt that the course was a success. Students indicated that they liked the course and thought that it was useful. The course also allowed students to express themselves creatively through the course project. The results were impressive and I’ll end this post with an example. The video below is the work of UNH ECE seniors Amy Schwarzenberg and Kyle Maroney (both graduated in May). Amy and Kyle explored user interactions with a Microsoft Surface multi-touch table.


[1] John Krumm (editor), “Ubiquitous Computing Fundamentals,” CRC Press, 2010

[2] Aaron Quigley, “From GUI to UUI: Interfaces for Ubiquitous Computing,” in John Krumm (editor), “Ubiquitous Computing Fundamentals,” CRC Press, 2010

[3] Mark Weiser, “The Computer for the 21st century,” Scientific American, pp. 94-10, September 1991

Return visit to Budapest University of Technology and Economics (BUTE)

On June 7 and 8, 2010 I visited the Budapest University of Technology and Economics (BUTE) for the second time in ten months. As with my last visit I went to discuss the BUTE-CEPS exchange program.

During this visit I met six people who have been involved in organizing different aspects of the exchange program. My host was Eszter Kiss, the Program Director of the Information Center for Engineering Programs in English (ICEPE). For UNH/CEPS students, staff and faculty, she is the Hungarian face of the exchange program. Eszter and I primarily talked about the fact that, starting in 2011, UNH ECE exchange students will spend the spring semester in Budapest. Other CEPS students will remain on the fall-in-Budapest schedule.

Eszter organized two meetings for me with BUTE leaders. The first one was with Dr. Peter Moson, Vice-Rector for International Relations (the Vice-Rector position at BUTE is equivalent to the Vice President position at a US university). Ildiko Varga, the head of the BUTE Erasmus and Exchange Office was also present at this meeting. Dr. Moson expressed his full support for a vibrant relationship between BUTE and CEPS. On a personal note it was great to see Dr. Moson who I met during his visit to UNH last year. It was also nice to talk to Ms. Varga who went to graduate school and taught mathematics at Purdue.

The second meeting organized by Eszter was with Dr. Gabor Stepan. Dr. Stepan, a member of the Hungarian Academy of Sciences (HAS), is the Dean of the BUTE Faculty of Mechanical Engineering, the ICEPE’s parent unit. Dr. Stepan expressed his full support for the BUTE-CEPS exchange program. Again on a personal note, it was exciting for me to visit the BUTE Faculty of ME where my father received his BS ME a long time ago. Dr. Stepan also spent some time telling me about BUTE’s history, including facts and anecdotes about BUTE’s Nobel-prize winning alumni.

While the meetings with Drs Stepan and Moson and with Ms. Varga primarily dealt with the overall BUTE-CEPS relationship, I also had a chance to work on issues related to UNH ECE directly with the BUTE unit that hosts ECE students. Specifically, Dr. Moson introduced me to Dr. Balint Kiss, the person in charge of the English language education at the Faculty of Electrical Engineering and Informatics. This is the BUTE unit that hosts UNH ECE exchange students and Dr. Kiss will be my primary contact in determining courses for our students to take while at BUTE. The meeting with Dr. Kiss was also an opportunity to catch up with Dr. Peter Arato. Dr. Arato, who is also a HAS member, has strong ties to the UNH ECE department having collaborated extensively with UNH ECE professor Andrzej Rucinski.

In addition to all these productive meetings I had a chance to give a talk to BUTE students interested in the exchage program. Seven prospective students attended, several of them interested in coming to the UNH ECE department – I hope we’ll see them here soon.

I would like to thank Eszter Kiss for organizing my visit (on very short notice). I would also like to thank the BUTE faculty, staff and students who took time to meet with me. Finally, I would like to acknowledge the UNH ECE Department and the CEPS Dean’s office who jointly funded this visit.

For pictures about my trips to Budapest visit my Flickr page.

Visit to FTW, Vienna

On June 4, 2010 I visited the Telecommunications Research Center Vienna (FTW). My host was Peter Froehlich, Senior Researcher in FTW’s User-Centered Interaction area of activity. Peter and I met at the CHI SIG meeting on automotive user interfaces [1] that I helped organize.

Peter and his colleagues are investigating automotive navigation aids and are currently preparing for an on-road study. I’m happy to report that this study will utilize one of our eye trackers. My visit provided an opportunity for us to discuss this upcoming study and how the eye tracker may be useful in evaluating the research hypotheses. Part of this discussion was a Telecommunications Forum talk I gave – see the slides below:

I want to thank Peter and his colleagues at FTW for hosting me and I’m looking forward to our upcoming collaboration. I also want to thank FTW for providing funding for my visit.


[1] Albrecht Schmidt, Anind L. Dey, Andrew L. Kun, Wolfgang Spiessl, “Automotive User Interfaces: Human Computer Interaction in the Car,” CHI 2010 Extended Abstracts

Albrecht Schmidt visit to UNH

Last month (April 16) Albrecht Schmidt visited UNH and the Project54 lab. Albrecht gave an excellent talk introducing some of the research problems in pervasive computing and specifically touching on the latest results from his lab, which were just published at CHI 2010 [1, 2]. I was especially interested in the work on helping users find the last place of interest on a map quickly. Albrecht and colleagues track the user’s gaze and when the user looks away, they place a marker (or gazemark) on the map. When the user looks back at the map she can start where she left off: at the place of the marker. Clearly this could be very useful when looking at GPS maps in a car. In such a situation the driver has to keep going back and forth between the map and the road and you want to minimize the time spent looking at the map (the road being the more important thing to look at!). The gazemarks introduced by Albrecht’s group may help. It would be interesting to conduct a driving simulator study with gazemarks.

After the talk Albrecht spent about an hour with students from the Project54 lab and those in my Ubicomp Fundamentals course. This was a more intimate setting for conversations about Albrecht’s research. Finally, Project54 staff and students spent a couple of hours discussing Project54 research with Albrecht – our work on handheld computers, on driving simulator-based investigations of in-car user interfaces and our budding efforts in multi-touch table interaction.

I am grateful to the UNH Provost’s Office for helping to fund Albrecht’s visit through a grant from the Class of 1954 Academic Enrichment Fund.


[1] Dagmar Kern, P. Marshall and Albrecht Schmidt, ” Gazemarks: gaze-based visual placeholders to ease attention switching,” CHI 2010

[2] Alireza Sahami Shirazi, Ari-Heikki Sarjanoja, Florian Alt, Albrecht Schmidt, and Jonna Häkkilä, J. “Understanding the impact of abstracted audio preview of SMS,” CHI 2010

MERL gift

I’m happy to report that I received a gift grant in the amount of $5,000 from Mitsubishi Electric Research Laboratories (MERL). The gift is intended to support my work on speech user interfaces and it was awarded by Dr. Kent Wittenburg, Vice President & Director of MERL.

This gift comes in the context of ongoing interactions between researchers at MERL and my group at UNH. Kent and Bent Schmidt-Nielsen hosted me several years ago for a demonstration of the Project54 system (I drove to Boston in a police SUV, which was fun), and I also gave a talk at MERL last fall.  In 2009 my PhD student Zeljko Medenica worked as a summer intern at MERL under the direction of Bret Harsham (Bret recently gave a talk at UNHon some of this work – see picture below). Zeljko is headed back to MERL this summer and he will work under the direction of Garrett Weinberg.

I greatly appreciate MERL’s generous gift and I plan to use it to help fund a graduate student working on speech user interfaces. I hope to report back to Kent, Bent, Bret and Garrett on the student’s progress by the end of this summer.

Project54 on front page of New York Times

In a front page article of the March 11, 2010 edition of the New York Times Matt Richtel discusses in-vehicle electronic devices used by first responders. Based on a number of interviews, including one with me, Matt gets the point across that interactions with in-vehicle devices can distract first responders from the primary task for any driver: driving. The personal accounts from first responders are certainly gripping. Thanks Matt for bringing this issue to the public.

Enter Project54. According to Matt “[r]esearchers are working to reduce the risk.” He goes on to describe UNH’s Project54 system which allows officers to issue voice commands in order to interact with in-car electronic devices. This means officers can keep their eyes on the road and their hands on the wheel. The article includes praise for the Project54 system by Captain John G. LeLacheur of the New Hampshire State Police. The Project54 system was developed in partnership with the NHSP and almost every NHSP cruiser has the Project54 system installed.

Both the print and the online versions of the article begin with a picture of the Project54 in-car system. This great picture was taken by Sheryl Senter and it shows Sergeant Tom Dronsfield of the Lee, NH Police Department in action.

Associate Professor, Electrical and Computer Engineering, University of New Hampshire