HCI Doctoral Consortium at VTT Oulu

Jonna Hakkila (Nokia), Jani Mantyjarvi (Nokia & VTT), and I discussed last year how we can improve the doctoral studies of our students and we decided to organize a small workshop to discuss PhD topics.

As Jonna is currently on maternity leave and officially not working we ran the workshop at VTT in Oulu.

The topics varied widely from basic user experience to user interface related security. There was very interesting work the participants did and published. I have selected the following 2 as reading suggestions: [1] by Elina Vartiainen and [2] by Anne Kaikkonen.

We hope we gave some advise – can resist to repeat the most important thing to remember:

  • a PhD thesis is not require to solve all problems in a domain
  • doing a PhD is yet another exam – not more and not less
  • finding/inventing/unterstanding something that makes a real difference to even a small part of the world is a great achievement (an not common in most PhD research)
  • do not start with thinking hard – start with doing your research

A good discussion on doing a PhD in computer science by Jakob Bardram can be found at [3].

[1] Roto, V., Popescu, A., Koivisto, A., and Vartiainen, E. 2006. Minimap: a web page visualization method for mobile phones. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (Montréal, Québec, Canada, April 22 – 27, 2006). CHI ’06. ACM, New York, NY, 35-44. DOI= http://doi.acm.org/10.1145/1124772.1124779

[2] Lehikoinen, J. T. and Kaikkonen, A. 2006. PePe field study: constructing meanings for locations in the context of mobile presence. In Proceedings of the 8th Conference on Human-Computer interaction with Mobile Devices and Services (Helsinki, Finland, September 12 – 15, 2006). MobileHCI ’06, vol. 159. ACM, New York, NY, 53-60. DOI= http://doi.acm.org/10.1145/1152215.1152228

[3] http://www.itu.dk/people/bardram/pmwiki/pmwiki.php?n=Main.ArtPhD

Christian Kray visits our Lab

Christian Kray and I were colleagues in Lancaster for a very short time – he just joined the university when I left for Munich. After his post-doc in Lancaster he moved to a position in Newcastle.

His work at the cross roads of mobile interaction and public displays is very exciting. In particular he investigates interesting concepts related to visual codes – some aspects to these ideas are discussed in “Swiss Army Knife meets Camera Phone” [1]. His new prototypes are really cool and I look forward to see/read more about them.

We realized that there are many areas where we have common interests. Perhaps there is a chance in the future to work together on some of the ideas discussed!

[1] Swiss Army Knife meets Camera Phone: Tool Selection and Interaction using Visual Markers. C. Kray and M. Rohs. (2007) In „Workshop on Mobile Interaction with the Real World at Mobile HCI 2007“. Singapore, September 9, 2007.

Matthias Kranz defended his PhD

Today Matthias Kranz defended his PhD thesis „Engineering Perceptive User Interfaces“ successfully at the University of Munich! Congratulations… and remember „Training to become a Jedi is not an easy challenge. And even if you succeed, it’s a hard life.“ (Qui-Gon Jinn, Star Wars Episode I: The Phantom Menace)

After Enrico Rukzio Matthias is the second of my PhD students who worked with me in Munich to finish. He is now with the German Aerospace center (DLR).
Matthias worked on the DFG project “Embedded Interaction” and has created and published an impressive set of prototypes and studies in pervasive computing over the last 4 years. See his homepage for details. If you only have time to read 2 pages – here is my suggestion: Context-aware kitchen utilities [1].
The personalized doctoral hat is a German tradition –and Matthias got a fully functional prototype of a perceptive doctoral hat (following the modification and restriction approach suggested in his thesis, but created without end-user involvement to keep the surprise 😉

Insider hint: http://en.wikipedia.org/wiki/Microfilm – just don’t!

[1] Kranz, M., Schmidt, A., Maldonado, A., Rusu, R. B., Beetz, M., Hörnler, B., and Rigoll, G. 2007. Context-aware kitchen utilities. In Proceedings of the 1st international Conference on Tangible and Embedded interaction (Baton Rouge, Louisiana, February 15 – 17, 2007). TEI ’07. ACM, New York, NY, 213-214. DOI= http://doi.acm.org/10.1145/1226969.1227013

Nicolas Villar visiting

Nicolas, who was the first BSc student I worked with in Lancaster, is now after finishing his PhD with Microsoft Research in Cambridge, UK. He came on Friday to Essen to see the lab and he brought us a Voodoo I/O box [1] – we are really excited!

He stayed for the weekend and I learned a lot about interesting technologies and ideas. Looking at his iREX ebook and Vivien’s new USB Microscope (30€ from Aldi 😉 we had to do some research into the screen quality of different devices. It is interesting to see that e-Ink moves closer to newspaper and that in comparison to it an iPhone screen is pretty coarse.

Some references to remember:

[1] Spiessl, W., Villar, N., Gellersen, H., and Schmidt, A. 2007. VoodooFlash: authoring across physical and digital form. In Proceedings of the 1st international Conference on Tangible and Embedded interaction (Baton Rouge, Louisiana, February 15 – 17, 2007). TEI ’07. ACM, New York, NY, 97-100. DOI= http://doi.acm.org/10.1145/1226969.1226989

Emmy Noether Meeting in Postdam

From Friday to Sunday I was in Potsdam for the yearly Emmy Noether meeting organized by DFG (German Science Foundation). The Emmy Noether Program seems to me one of the most attractive funding options for early career researchers I know of (world wide).

This year I was in the preparation team for the meeting and was co-organizing one workshop where we discussed experiences with research funding in other countries and what ideas for improving current programs we can see. The workshop was on how we can benefit from having students rather than seeing them as “teaching load” – especially in computer science (I organized it with Andreas Butz).

The meeting is always very interesting as it brings together people, funded by DFG in the Emmy Noether Program, across all disciplines. In the political evening we had a keynote by Professor Hans Weiler looking at the current problems of the German systems („Eliten im Wettbewerb – Die deutschen Hochschulen und die internationale Konkurrenz“ ) – one message I saw in the talk is that Germany is very efficient – looking how little money is spend in education and research the outcome is surprising. (But this is only a positive message if we do not want to play a leading role in the world of science and technology). It became very clear that the overall system lacks massively in funding. The additional funding that is provided by the German Government in the widely publicized call for elite universities (Exzellenzinitiative) is 1900 million Euros over 5 years (about 5€/citizen/year)– impressive? Not really – this less than the amount projected for the “Yale Tomorrow” campaign – a 5 year program in fundraising by a single University in the US. And Stanford University has even a bigger campaign as Prof. Weiler told – and there are a few other Universities in the league in the US…

GPS monitoring for car insurance

In my talk at ISUVR2008 I referred to an example where an insurance is monitor driving behavior and makes a tariff according to this. Some people asked me for more details and
references, here they are…

My example was based on the pilot announced from the German insurance WGV. They planned to run a pilot with 1500 people using a GPS based monitoring devices. The box is mounted in the car and compares the current speed with the allowed speed limit and warns to reduce speed (if over the limit). If the driver is more than 12 times per year over the speed limit (basically ignoring the warning) he does not get the reduced rate. (see http://www.wgv-online.de/docs/youngandsafe.pdf – in German only). In the announcement it said they will run the pilot to 2009…

There are different ideas how to take GPS driving monitoring beyond the lab, e.g. in 2007 Royal & SunAlliance announced a GPS-based eco car insurance and the AIG a Teen GPS Program – targeted at parents.

Looking at different comments (on news pages and in blogs) it seems that people’s opinions are very split…

Thermo-imaging camera at the border – useful for Context-Awareness?

When we re-entered South Korea I saw guard looking with an infrared camera at all arriving people. It was very hot outside so the heads were very red. My assumption is that this is used to spot people who have fever – however I could not verify this.

Looking at the images created while people moved around I realized that for many tasks in activity recognition, home health care, and wellness this may be an interesting technology to use. For several tasks in context-awareness it seems straightforward to get this information from an infrared camera. In the computer vision domain it seems that there have several papers towards this problem over the recent years.

We could thing of an interesting project topic related to infrared activity recognition or interaction to be integrated in our new lab… There are probably some fairly cheep thermo-sensing cameras around to used in research – for home brew use you find hints on the internet, e.g. How to turn a digital camera into an IR cam – pretty similar to what we did with the web cams for our multi-touch table.

The photo is from http://en.wikipedia.org/wiki/Thermography

Trip to North Korea

[see the whole set of photos from tour to North Korea]

From Gwangju we took the bus shortly after midnight to go for a trip to North Korea. The students did a great job in organizing ISUVR and the trip. It was great to have again some time to talk to Yoosoo Oh, who was a visiting researcher in Munich in our group.

When entering North Korea there are many rules, including that you are not allowed to take cameras with tele-lenses over 160mm (so I had to take only the 50mm lens) and you must not bring mobile phones and mp3 players with you. Currently cameras, phones and MP3 players are visible with the human eye and to detect in an x-ray. But it does not take much imagination to see in a few years extremely small devices that are close to impossible to spot. I wonder how this will change such security precautions and whether or not I will in 10 years still possible to isolate a country from access to information. I doubt it…

The sightseeing was magnificent – see the photos of the tour for yourself. We went onto the Kaesong tour (see http://www.ikaesong.com/ – in Korea only) It is hard to tell how much of the real North Korea we really saw. And the photos only reflect a positive selection of motives (leaving out soldiers, people in town, ordinary buildings, etc. as it is explicitly forbidden to take photos of those). I was really surprise when leaving the country they check ALL the pictures you took (in my case it took a little longer as it was 350 photos).

The towns and villages are completely different from what I have seen so far. No cars (besides police/emergency services/army/tourist busses) – but many people in the street walking or cycling. There were some buses in a yard but I have not seen public transport in operation. It seemed the convoy of 14 tourist buses is an attraction to the local people…

I have learned that the first metal movable type is from Korea – about 200 years before Gutenberg. Such a metal type is exhibited in North Korea and in the display is a magnifying glass in front of the letter – pretty hard to take a picture of…

ISUVR 2008, program day2

Norbert Streitz – Trade-off for creating smartness

Norbert gave an interesting overview of research in the domain of ubicomp based on his personal experience – from Xerox PARC to the disappearing computer. He motivated the transition from Information Design to Experience Design. Throughout the work we see a trade-off between providing “smart support” to the user and “privacy” (or control over privacy). One of the questions if we will re-invent privacy or if it will become a commodity…
As one of the concrete examples Norbert introduced the Hello.Wall done in the context Ambient Agoras [1]. This again brought up the discussion of public vs. private with regard to the patterns that are displays. (photos of some slides from Norbert’s talk)

[1] Prante, T., Stenzel, R., Röcker, C., Streitz, N., and Magerkurth, C. 2004. Ambient agoras: InfoRiver, SIAM, Hello.Wall. In CHI ’04 Extended Abstracts on Human Factors in Computing Systems (Vienna, Austria, April 24 – 29, 2004). CHI ’04. ACM, New York, NY, 763-764. DOI= http://doi.acm.org/10.1145/985921.985924 (Video Hello.Wall)
Albrecht Schmidt – Magic Beyond the Screen
I gave a talk on “Human Interaction in Ubicomp -Magic beyond the screen” highlighting work in user interfaces beyond the screen that we did over the last years. It is motivated by the facts that classical limitations in computer science (e.g. frame rate, processing, storage) are getting less and less important to many application areas and that the human computer interaction becomes in many areas the critical part of the system.
In my talk I suggested using “user illusion” as a design tool for user interfaces beyond the desktop. This involves two steps: 1) describe precisely the user illusion the application will create and the 2) Investigate what parameters have an influence on the quality of the created user illusion for the application. (photos of some slides from Albrecht’s talk, Slides in PDF)
Jonathan Gratch – Agents with Emotions

His talk focused on the domain of virtual reality with a focus on learning/training applications. One central thing I learned is that the timing of non-verbal cues (e.g. nodding) is very crucial to produce an engagement in speaking with an agent. This may also be interesting for other forms of computer created feedback.
He gave a specific example on how assigning blame works. It was really interesting to see that there are solid theories in this domain that can be concretely used to design novel interfaces. He argues that appraisal theory can explain people’s emotional states and this could improve context-awareness.

He showed an example of emotional dynamics and it is amazing how fast emotion happen. One of the ways of explaining this is to look at different dynamics: dynamics in the world, dynamics in the perceived world relationship, and dynamic through action. (photos of some slides from Jonathan’s talk)
Daijin Kim – Vision based human robot interaction
Motivated by the vision that after the personal computer we will see the “Personal Robot” Daijin investigates natural ways to interact with robots. For vision based interaction with robots he named a set of difficulties, in particular: people are moving, robots are moving, and the illuminations and distances are variable. The proposed approach is to generate a pose, expression, and illumination specific active appearance model.
He argues that face detection is a basic requirement for vision based human robot interaction. The examples he showed in demo movie were very robust with regard to movement, rotation, and expression and it works for very variable distances. The talk contained further examples of fast face recognition and recognition of simple head gestures. Related to our research it seems that such algorithms could be really interesting in creating context-aware outdoor advertisement. (photos of some slides from Daijin’s talk)

Steven Feiner – AR for prototyping UIs

Steven showed some work mobile projector and mobile device interaction, were they used augmented reality for prototyping different interaction methods. He introduced Spot-light (position based interaction), orientation based interaction and widget-based interaction for an arm mounted projector. Using the synaptic touchpad and projection may also be an option for our car-ui related research. For interaction with a wrist device (e.g. a watch) he introduced the string-based interaction which is a simple but exciting idea. You pull out a string of a device and the distances as well as the direction are the resulting input parameters [2].
In a further example Steven showed a project that supports field work on identification of plants using capture (of the image of the real leaf), comparison with the data base and matching out of a subset that matches the features. Their prototype was done on a tablet and he showed ideas how to improve this with AR; it is very clear that this may also an interesting application (for the general user) on the mobile phone.

New interfaces and in particular gestures are hard to explore – if you have no idea what is supported by the system. In his example on visual hint for tangible gestures using AR Steven showed interesting options in this domain. One approach follows a “preview style” visualizations – they called it ghosting. (photos of some slides from Stevens’s talk)

[2] Blasko, G., Narayanaswami, C., and Feiner, S. 2006. Prototyping retractable string-based interaction techniques for dual-display mobile devices. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (Montréal, Québec, Canada, April 22 – 27, 2006). R. Grinter, T. Rodden, P. Aoki, E. Cutrell, R. Jeffries, and G. Olson, Eds. CHI ’06. ACM, New York, NY, 369-372. DOI= http://doi.acm.org/10.1145/1124772.1124827
[3] White, S., Lister, L., and Feiner, S.Visual Hints for Tangible Gestures in Augmented Reality.Proc. ISMAR 2007 IEEE and ACM Int. Symp. on Mixed and Augmented Reality, Nara Japan, November 13-16, 2007. (youtube video)

If you are curious about the best papers, please the photos from the closing 🙂

Finally some random things to remember:

  • Richard W. DeVaul did some work on subliminal user interfaces – working towwrds the vision of zero attention UIs [4]
  • Jacqueline Nadel (development psychologist) did studies on emotions between parents and infants using video conferencing
  • V2 – Toward a Universal Remote Console Standard http://myurc.org/whitepaper.php
  • iCat and Gaze [5]

[4] Richard W. DeVaul. The Memory Glasses: Wearable Computing for Just-in-Time Memory Support. PhD Thesis. MIT 2004. http://devaul.net/~rich/DeVaulDissertation.pdf

[5] Poel, M., Breemen, A.v., Nijholt, A., Heylen, D.K., & Meulemans, M. (2007). Gaze behavior, believability, likability and the iCat. Proceedings Sixth Workshop on Social Intelligence Design: CTIT Workshop Proceedings Series (pp. 109–124). http://www.vf.utwente.nl/~anijholt/artikelen/sid2007-1.pdf

Korean Dinner – to many dishes to count

In the evening we had a great Korean dinner. I enjoyed it very much – and I imagine we have seen everything people eat in Korea – at some point I lost count of the number of different dishes. The things I tasted were very delicious but completly different to what I typically eat.

Dongpyo Hong convinced me to try a traditional dish (pork, fish and Kimchi) and it was very different in taste. I was not adventures enough to try a dish that still moved (even though the movement was mariginal – can you spot the difference in the picture) – but probably I missed something as Dongpyo Hong enjoyed it.

I made some photos from the conference dinner.