Redesign of Slovenian Avalanche Bulletin

Vanja Blažica, Janez Jaka Cerar, Aleš Poredoš
Slovenian Environment Agency, Vojkova 1b, 1000 Ljubljana, Slovenia

Abstract: We present the redesign of the Slovenian avalanche bulletin, published regularly during the winter season to warn against avalanche danger and to provide specific information for advanced users. The former version included an estimation of danger on a scale from one to five with supporting text for the whole country, while the new one offers an additional graphical description, specified for several geographical regions. The redesign profoundly influenced the work of avalanche forecasters by introducing a new interface, additional input and database storage. At the same time, users welcomed the additional information, international comparability and user friendliness of the new bulletin.


Full paper: link

Designing visual interface for nutrition tracking of patients with Parkinson’s disease

Peter Novak, Barbara Koroušić Seljak, Franc Novak
Jožef Stefan Institute, Jamova cesta 39, 1000 Ljubljana, Slovenia

Abstract: In this paper, we describe the design of a visual interface of a mobile a mobile app for tracking nutrients and foods consumed by patients with Parkinson’s disease. The interface should enable the patients to recognize objects on the screen, easily perceive their function and interact with them thus providing an efficient way of entering the dietary intake data. The app has been validated by five patients and the preliminary results are encouraging.


Full paper: link


3D serious games for Parkinson’s disease management

Bojan Blažica, Franc Novak, Anton Biasizzo
Jožef Stefan Institute, Jamova cesta 39, 1000 Ljubljana, Slovenia
Ciril Bohak
Faculty for Computer and Information Science, University of Ljubljana, Večna pot 113, 1000 Ljubljana, Slovenia

Abstract: The aim of the article is to show how off-the-shelve equipment can be used to develop serious games for an affordable tele-medicine solution for Parkinson’s disease management. Two games have been developed aimed at assessing and training patient’s reach of upper limbs (using Kinect v2) and fine motoric skills of fingers (using Leap motion). The games collect player data in terms of score achieved and full kinematics of movement during gameplay. The data is stored online and made available to therapists and doctors through a secure connection. The games have been tested with patients within the Soča rehabilitation institute as well as at their homes.


Full paper: link


Remote Interaction in Web-Based Medical Visual Application

Ciril Bohak, Primož Lavrič, Matija Marolt
Faculty for Computer and Information Science, University of Ljubljana, Večna pot 113, 1000 Ljubljana, Slovenia

Abstract: In this paper we present a novel integration of four remote collaboration modalities into an existing web-based medical data visualization framework: (1) visualization data haring, (2) camera view sharing, (3) data annotation haring and (4) chat. The integration of remote collaboration modalities was done for two reasons: for getting the second opinion on diagnosis or for getting a diagnosis from the remote medical specialist. We present an integration of these modalities and a preliminary evaluation by the medical expert. In conclusion we show that we are on the correct track of integrating collaboration modalities into the visualization framework.


Full paper: link

HCI-IS 2016 Schedule

Part 1 (14.00 – 16.00)

  • 14.00 – 14.15 – Opening event
  • 14.15 – 14.30 – Remote Interaction in Web-Based Medical Visual Application (Ciril Bohak, Primož Lavrič, Matija Marolt)
  • 14.30 – 14.45 – 3D serious games for Parkinson’s disease management (Bojan Blažica, Franc Novak, Anton Biasizzo, Ciril Bohak)
  • 14.45 – 15.00 – Designing visual interface for nutrition tracking of patients with Parkinson’s disease (Peter Novak, Barbara Koroušić Seljak, Franc Novak)
  • 15.00 – 15.15 – Short break
  • 15.15 – 15.30 – Redesign of Slovenian Avalanche Bulletin (Vanja Blažica, Janez Jaka Cerar, Aleš Poredoš)
  • 15.30 – 15.45 – Improving the usability of online usability surveys with an interactive Stripe scale (Matevž Pesek, Alja Isakovič, Gregor Strle, Matija Marolt)
  • 15.45 – 16.00 – Evaluation of common input devices for web browsing: mouse vs touchpad vs. touchscreen (Andrej Malečkar, Matjaž Kljun, Klen Čopič Pucihar, Peter Rogelj)

Coffee break (16.00 – 16.30)

Part 2 (16.00 – 17.45)

  • 16.30 – 16.45 – Wizard of Oz experiment for Prototyping Multimodal Interfaces in Virtual Reality (Blaž Gombač, Damir Deželjin, Matej Zemljak, Klen Čopič Pucihar, Patrik Širol, Matjaž Kljun)
  • 16.45 – 17.00 – Towards the improvement of GUARD graphical user interface (Žiga Kopušar, Franc Novak)
  • 17.00 – 17.15 – Towards affordable Mobile Crowd Sensing device (Gal Pavlin, Marko Pavlin)
  • 17.15 – 17.30 – I was here: a system for creating augmented reality digital graffiti in public place (Erik Šimer, Matjaž Kljun, Klen Čipič Pucihar)
  • 17.30 – 17.45 – Interactive Video Management by means of an Exercise Bike (Jan Štrekelj, Branko Kavšek)

Conference closing

TETRACOM project: Personalized Nutrition Control Aid for Insulin Patch Pump – PerNuCAP

Project partners: Jozef Stefan Institute, Ljubljana and IPD Med d.o.o., Šentjernej

Diabetes is a chronic condition marked by high levels of glucose in the blood. In type 1 diabetes, the body completely stops producing any insulin, a hormone that enables the body to use glucose found in foods for energy. People with type 1 diabetes must daily receive insulin to survive. Insulin pump therapy offers efficient metabolic disorder control and represents a cost effective alternative to multiple daily insulin injections. Compared with ‘classical’ pumps, insulin patch pumps offer patients more freedom. They are the ‘next generation’ devices: discreet, smaller (the size of the patch) and a lot more convenient to wear. They allow patients to engage in everyday activities such as sport and even swimming.

The Insulin Patch Pump developed by IPD Med is designed for continuous insulin delivery according to personalized, pre-programmed plan. The device provides both basal and bolus modes of insulin delivery. It consists of a reusable shell that attaches to a disposable unit. The device is attached to a body with an adhesive patch. The shell contains a battery powered microprocessor with corresponding controls, alerting devices and RF circuitry. The disposable unit includes a pumping mechanism, proprietary insulin reservoir and COTS/OEM infusion set. An RF remote control unit is provided to manage and program the device.

The disposable unit represents only 0.5% of the total price, which is a considerable advantage in comparison with existing devices on the market. However, even more significant step forward is expected by providing the users with the possibility to access OPEN Platform For Clinical Nutrition developed by Jozef Stefan Institute (JSI). OPEN is a web-based application that supports food and physical activity recording and diet planning. OPEN enables online interaction between a dietitian and his (her) patient. To support its use in different countries, OPEN allows translation of the user interface into other languages as well as the use of any food composition dataset that complies with Food data structure and format standard (BS EN 16104:2012). OPEN refers to international, evidence-based dietary recommendations, which can be modified by the dietitian to suit the needs of individuals.

In connection with OPEN platform, the Personalized Nutrition Control Aid for Insulin Patch Pump will provide the user simple means to obtain detailed data on his/her nutrition intake needed for the prediction of insulin dosing profile. Personalized user interface will help the user to determine carbon hydrate intake with higher accuracy than currently established methods. Consequently, the insulin therapy is improved by reducing the risk of hypoglycemia caused by improper data input. The aid will also allow the user to keep the food diary and provide access to extensive food lexicon. Personalized user interface is currently under construction. Its concept closely resembles a similar one designed as a mobile app in the frame of PD_Manager, which aims of collecting data about nutrition in patients with Parkinson’s disease.

IPDmed insulin patch pump and OPEN platform

Since the designed Insulin Patch Pump is classified as class IIb medical device according to MDD 93/42/EEC, in-depth risk management including usability testing must be performed. In this regard, extensive risk analysis has been performed by IPD Med, while JSI, who has proven expertise in human-computer interaction has prepared the assessment usability test plan in accordance with Standard IEC 62366 (Medical devices – Application of usability engineering to medical devices). Usability testing with a target group of volunteers is planed for September. While work on the project is still running and in accordance with the contract between JSI and IPD Med will continue until the end of October 2016, the current results of the TTP supported by TETRACOM represent an important step toward TRL 7.

View on the Tetracom project website or download poster here.

Natural User Interfaces – NUI

“Computing machines can do readily, well, and rapidly many things that are difficult or impossible for men and men can do readily and well, though not rapidly, many things that are difficult or impossible for computers. This suggests that a symbiotic cooperation, if successful in integrating the positive characteristics of men and computers, would be of great value” (Licklider, 1960).

In 1960, Licklider identified the problems for realization of such ‘man-computer symbiosis’ in speed mismatch between men and computer, memory hardware requirements, memory organization requirements, differences between human and computer language and input and output equipment, the latest being the most problematic.

Fifty years later this still holds true.

NUI - importance of reusing skills

Word cloud based on this post about NUIs. The most emphasized
word, ‘skills’ highlights the importance of making NUIs feel intuitive to use, which can
be achieved by designing the interface so that it relies on skills that users already possess

Today Human-Computer Interaction (HCI) is concerned with the above-mentioned problems i.e. the design, evaluation and implementation of interactive computing systems for human use and the study of major phenomena surrounding them (Hewett et al., 2009). HCI is an interdisciplinary area deriving its knowledge from computer science (application design and engineering of human interfaces), psychology (the application of theories of cognitive processes and the empirical analysis of user behaviour), sociology and anthropology (interactions between technology, work, and organization), and industrial design (interactive products).

Interaction between human and computers takes place at the user interface. This is why research and development of user interfaces lies at the very core of HCI. We will first briefly present the development that led to today’s user interfaces and then focus on multitouch displays, their construction and use cases.

From Punch Cards to Natural User Interfaces

Early digital computers used batch interfaces. They consist of punch cards for programs and data input and prints as output. Batch interfaces are non-interactive, which means that the user has to specify everything prior to processing and is never prompted for additional input until the job is done. With the connection of teletype machines to computers and the advent of dedicated text-based CRT terminals in the late 50s, batch interfaces gave space to command line interfaces (CLI). CLIs brought speed and interactivity to human computer interaction. In the early 60s, Douglas Engelbart invented the first mouse prototype used for manipulating text-based hyper-links as a part of his ‘augmenting human intellect’ project (Engelbart, 1962). Later, researchers at Xerox PARC extended the concept of hyper-links to graphics and by doing so created the first graphical user interface (GUI). In respect to the CLI, GUI facilitates a more intuitive interaction because there is no need to learn commands by heart as available commands are presented on the screen in the form of windows, icons, menus and a pointer. This interaction paradigm can be emphasized by the acronym WIMP. Further efforts focused on the development of interfaces that interact with humans in a seamless way by understanding natural expressions of the user’s intent instead of forcing him/her to learn new rules for interaction. The results of these efforts are, among others, various speech and gesture recognition techniques known as Natural User Interfaces or NUIs.

Natural User Interfaces

NUI Group, a global research community focused on the open discovery of natural user interfaces, defines natural user interfaces as an emerging computer interaction methodology which focuses on human abilities such as touch, vision, voice, motion and higher cognitive functions such as expression, perception and recall. A natural user interface seeks to harness the power of a much wider breadth of communication modalities which leverage skills people gain through traditional physical interaction (NUI Group Community, 2009). Some representative examples of such interfaces are:

  • Multitouch displays (Han, 2005): tracking and recognizing multiple fingers (and objects) on a display, which leads to new interaction techniques. These displays differ from touchscreens capable of detecting one touch in that they allow for a different interaction paradigm, while touchscreens are just a replacement for the input device, usually the mouse, in a GUI.
  • Speech recognition (Rosenfeld et al., 2000): interfaces based on speech recognition, speech synthesis and natural language processing.
  • Space gesture recognition (Hay et al., 2008): 3D tracking or motion capture using Nintendo Wii or similar controllers as stereo vision systems. The user interacts with the system by performing, usually predefined, spatial gestures.
  • Interfaces based on electrophysiological signals (e.g. Brain-Computer Interfaces, Wolpaw (2002)): these interfaces determine the intent of the user from electrophysiological signals caused by brain, muscle, and cerebral cortex activity. For example, classifying finger gestures by interpreting forearm electromyography (EMG) signals caused by muscle movements (Saponas et al., 2009). These interfaces are also known as direct neural interfaces, mind-machine interfaces or brain-machine interfaces.

The main property of these interfaces is their intuitiveness. Ideally, a natural user interface does not require the user to undergo any training in order to interact with it. In other words, NUIs excel in terms of learnability and discoverability. NUIs build on the knowledge users get in their everyday life and exploit interaction metaphors with real-world objects to interact with digital objects in the digital world. The actual definition of NUIs is subject of ongoing debate and there are many definitions available. Some of the more concise definitions are:

  • TechTarget: “A natural user interface (NUI) is a system for human-computer interaction that the user operates through intuitive actions related to natural, everyday human behaviour … A NUI may be operated in a number of different ways, depending on the purpose and user requirements. Some NUIs rely on intermediary devices for interaction but more advanced NUIs are either invisible to the user or so unobtrusive that they quickly seem invisible.”
  • TechTerms: “A NUI is a type of user interface that is designed to feel as natural as possible to the user. The goal of a NUI is to create seamless interaction between the human and machine, making the interface itself seem to disappear.”
  • Wikipedia, Ron George: “…a user interface that is effectively invisible, or becomes invisible with successive learned interactions, to its users.”

Some authors offer a more lengthy definition of natural user interfaces or give a short definition accompanied by a lengthier explanation, like Joshua Blake’s definition of NUIs:

“A natural user interface is a user interface designed to reuse existing skills for interacting directly with content (Blake, 2013).”

He explains his definition as follows:

NUIs are designed. First, this definition tells us that natural user interfaces are designed, which means they require forethought and specific planning efforts in advance. Special care is required to make sure NUI interactions are appropriate for the user, the content, and the context. Nothing about NUIs should be thrown together or assembled haphazardly. We should acknowledge the role that designers have to play in creating NUI style interactions and make sure that the design process is given just as much priority as development;

NUIs reuse existing skills. Second, the phrase ‘reuse existing skills’ helps us focus on how to create interfaces that are natural. Your users are experts in many skills that they have gained just because they are human. They have been practicing for years skills for human-human communication, both verbal and non-verbal, and human-environmental interaction. Computing power and input technology has progressed to a point where we can take advantage of these existing non-computing skills. NUIs do this by letting users interact with computers using intuitive actions such as touching, gesturing, and talking and presenting interfaces that users can understand primarily through metaphors that draw from real-world experiences. This is in contrast to GUI, which uses artificial interface elements such as windows, menus, and icons for output and pointing device such as a mouse for input, or the CLI, which is described as having text output and text input using a keyboard. At first glance, the primary difference between these definitions is the input modality – keyboard versus mouse versus touch. There is another subtle yet important difference: CLI and GUI are defined explicitly in terms of the input device, while NUI is defined in terms of the interaction style. Any type of interface technology can be used with NUI as long as the style of interaction focuses on reusing existing skills;

NUIs have direct interaction with content. Finally, think again about GUI, which by definition uses windows, menus, and icons as the primary interface elements. In contrast, the phrase ‘interacting directly with content’ tells us that the focus of the interactions is on the content and directly interacting with it. This doesn’t mean that the interface cannot have controls such as buttons or checkboxes when necessary. It only means that the controls should be secondary to the content, and direct manipulation of the content should be the primary interaction method.”

In his keynote speech at Interact 2011, Antão Almada emphasizes Blake’s ‘interaction style over input modality’ point of view: “People feel naturally what they are supposed to do. Natural means ease of use, non-invasive sensors, to simplify as much as possible, to give intelligence to the interface so that the users get only the useful information.” The company he works for uses NUIs mainly in marketing applications, where people do not have time to learn how to interact with the applications they build and the need for easy, learnable interaction based on skills people already have is particularly strong.

Richard Monson-Haefel’s presents yet another variant of NUI definition: “A Natural User Interface is a human-computer interface that models interactions between people and the natural environment.” He continues specifying a couple of aspects of this definition that are especially important:

“NUI is a form of HCI. It is important that we make that explicit;

NUI models natural interactions. That means it leverages and uses as a template the interactions people have with each other (e.g. speech and gestures); NUI also models interactions between people and the natural environment (e.g. water and rocks) as opposed to their artificial environment (e.g. computers and cars.).”

Dennis Wixon talks about NUIs and interfaces in general in terms of principals and guidelines (Wixon, 2008). He believes that from a historical analysis of user interfaces we can extract a set of principles and guidelines for each interface type. “The principles are what drives the design. The guidelines are simply derivations of the principles for an individual context. Principles are what’s important. Principles and data drive successful design.” In this sense he analyses three different interface types that primarily build upon text—command line interfaces, graphics—graphical user interfaces and objects—natural user interfaces. He notes that interaction in CLIs builds upon the psychological function of recall as the user is disconnected from the static system he interacts with and has to direct the system by learning/recalling a high number of commands. He continues by describing GUIs as an exploratory and responsive type of interaction, where the user scans through menus and recognizes the commands he had to learn and recall within CLIs. Wixon argues that interaction with GUIs is indirect as the user actually controls the mouse or the keyboard and not the GUI itself. NUIs, on the other hand, provide an unmediated interaction that is evocative and thus relies on user intuition. The commands in a NUI are few and the interaction is fast. Wixon states that NUIs are also contextual and that they understand the environment they are in and react to it naturally. He concludes this analysis of interfaces by extracting a set of principles that should apply to NUI design: the principal of performance aesthetics (the interaction should be enjoyable), the principle of direct manipulation, the principle of scaffolding (the system should support actions as you move forward and should reveal itself in those actions), the principle of contextual environments and the principal or the super real (to take real things and extend them in a logical yet unrealistic way in the digital world). These 5 principles build upon three core principles: social, seamless, spatial. These are explained on an example – the Microsoft Surface multitouch tabletop computer. Surface is social as it encourages social interaction around a shared information space and brings people together instead of isolating them, the actions of a Surface user and the interaction with the Surface are the same, which makes the Surface seamless and spatial as the objects in the interface have an implied physicality and the interaction with the Surface leverages spatial memory.

Wixon co-authored a book about NUIs with Daniel Wigdor, called ‘Brave NUI world’ (Wigdor and Wixon, 2011). The book presents guidelines stemming from the abovementioned principles and tries to define NUIs from a user centred perspective: “The word ‘natural’ in natural user interfaces describes the user’s feelings during interaction and not the interface itself. Interfaces often cited as NUIs, e.g. multitouch displays with their highbandwidth input modality, are not natural per se – they instead provide a higher potential for developing a natural user interface, if and only if an all-new interface is designed with new input actions, new affordances – in short, a new paradigm.” In their view, creating a natural user interface is a design goal. It can be achieved through “a clear viewpoint, hard work, careful design, rigorous testing and some luck.” The clear viewpoint here is their vision of natural user interfaces: “Our vision is that a natural user interface is one that provides a clear and enjoyable path to unreflective expertise in its use. It makes skilled behaviour seem natural in both learning and expert practice. It makes learning enjoyable and eliminates the drudgery that distracts from skilled practice. It can make you a skilled practitioner who enjoys what you are doing. Natural in this sense does not mean raw, primitive, or dumbed down.” Finally, they summarize their definition of natural user interfaces as follows:

“A NUI is not a natural user interface, but rather an interface that makes your user act and feel like a natural. An easy way of remembering this is to change the way you say ‘natural user interface’ – it’s not a natural user interface, but rather a natural user interface.”

Bill Buxton supports this natural user interface over natural user interface point of view: “It’s not about speech, it’s not about gesture, it’s not even about the phone, and it’s not about human-to-human communication,” he says. “How these things work together in a natural and seamless way that reduces complexity for the users — that’s what we’re about. Getting these things right opens up another dimension in how we have technology integrated into our lives” (Buxton, 2010). His understanding of what natural in natural user interfaces means is: “designing to take advantage of the skills we acquired in a lifetime of living in the real world. These skills are motor-sensor skills, cognitive skills and social skills.” In this sense, a designer of natural user interfaces must take care not to waste people’s skills. This design paradigm also leads to interfaces that are natural to some users and not natural to others. For example, someone that has invested the time and effort to learn touch typing may find editing text in ‘vi’ natural. On the other hand, despite touch being usually regarded as a natural means for interaction, Foehrenbach et al. (2008) surprisingly report how tactile feedback added to gestural interaction with high resolution interactive walls increases error rates by 10 %. Kurfess (2013) shares this hardware-agnostic point of view on NUIs. In his opinion, natural user interfaces are not about input modalities, but about interaction style. A user interface becomes natural with careful interaction design and planning; the word natural here means that interaction is appropriate for the user, the content, and the context in which the interaction takes place. This can be achieved by reusing skills that the user already has and building upon experience and expertise often unrelated to computer use. Kurfesses key guidelines for developing natural user interfaces are the use of direct manipulation where possible, enabling instant expertise for the user, reducing cognitive load and inducing progressive learning.

The roots of this debate about the essence of natural interaction can be traced back to the late 70s, and early 80’s. Although at the time the expression natural user interfaces was not yet born, Ben Shneidermann’s work on direct manipulation belongs in this context. In (Frohlich, 1993) Frohlich et al. summarized direct manipulation as “a style of interaction characterized by the following three properties:

  1. Continuous representation of the object of interest,
  2. Physical actions or labelled button presses instead of complex syntax, and
  3. Rapid incremental reversible operations whose impact on the object of interest is immediately visible.”

The benefits of direct manipulation are “ease of learning, ease of use, retention of learning, reduction and ease of error correction, reduction of anxiety and greater system comprehension” ( Shneiderman (1982), p. 253). Due to the overlapping of these benefits and the various definitions of natural user interfaces and their goals we can say that findings related to direct manipulation can be applied to NUIs as well. In 1985, Hutchins et al. (1985) deconstruct the term directness in direct manipulation in two parts: the psychological distance between user goals and the action a specific interface requires to achieve these goals, and to psychological engagement of feeling oneself to be controlling the computer directly rather than through some intermediary. “Essentially distance refers to the mismatch between the way a user normally thinks about a problem domain and the way it is represented by a computer. Systems which reduce distance reduce this mismatch and the associated mental effort of working out what can be done in the system (semantic distance) and how to do it (articulatory distance). Engagement, on the other hand, refers to a particular style of representation based on a model world metaphor rather than on a conversational metaphor for interaction. Systems can encourage a feeling of engagement by depicting objects of interest graphically and allowing users to manipulate them physically rather than by instruction.”( Frohlich (1993), p. 3) NUIs’ strategy to reduce the above-mentioned psychological distance between user goals and action is to exploit skills users already possess. It is this ‘recycling/remixing/reusing of skills’ that makes NUIs intuitive. However, a lower distance between goals and actions is not always good – to accomplish more abstract tasks, this distance is actually beneficial, for example: repetitive tasks are easier to accomplish within a command line interface with a ‘for loop’ than manually in a more direct user interface. In this sense, Hutchins et al. conclude that direct manipulation systems benefit from the simplified mapping between goals and actions needed in an interface to achieve those goals, but at the same time lose some expressive power due to the loss in abstraction corresponding to the simplified mapping. As Bill Buxton often puts it: “Everything is best for something and worst for something else.” Therefore in 1993 Frolich (Frohlich, 1993) proposed a shift in interaction from directness to gracefulness: “The puzzle of desirable indirectness in interaction is solved if we shift to a more social definition of directness as interaction in which there is least collaborative effort expended to achieve a users’ goals. Activities which are cognitively indirect can then be seen as socially direct in that they have the effect of minimizing the joint work carried out by system and user entailed in achieving task success.” Frolich’s notion of graceful interaction builds upon Hayes’ and Reddy’s (Hayes and Reddy, 1983), who in 1983 proposed a decomposition of the term graceful interaction into a set of skills: “skills involved in parsing elliptical, fragmented, and otherwise ungrammatical input; in ensuring robust communication; in explaining abilities and limitations, actions and the motives behind them; in keeping track of the focus of attention of a dialogue; in identifying things from descriptions, even if ambiguous or unsatisfiable; and in describing things in terms appropriate for the context.” We can say that graceful interaction addresses some of the generally acknowledged problems of NUIs. Don Norman in (Norman, 2010) argues that a pure gestural system makes it difficult to discover the set of possibilities and the precise dynamics of execution that an interface requires: “It is also unlikely that complex systems could be controlled solely by body gestures because the subtleties of action are too complex to be handled by actions – it is as if our spoken language consisted solely of verbs. We need ways of specifying scope, range, temporal order, and conditional dependencies. As a result, most complex systems for gesture also provide switches, hand-held devices, gloves, spoken command languages, or even good old-fashioned keyboards to add more specificity and precision to the commands. Gestural systems are no different from any other form of interaction. They need to follow the basic rules of interaction design, which means welldefined modes of expression, a clear conceptual model of the way they interact with the system, their consequences, and means of navigating unintended consequences.” He continues: “Whether it is speech, gesture, or the tapping of the body’s electrical signals for ‘thought control,’ all have great potential for enhancing our interactions, especially where the traditional methods are inappropriate or inconvenient. But they are not a panacea.”

If direct and graceful manipulation have been terms used to discuss NUI-like interaction in the past, some authors suggest new terms for future discussion. Among them are Oviatt and Cohen’s Perceptual User Interfaces (PUI) (Oviatt and Cohen, 2000), Wixon’s Organic User Interfaces (OUI) (Wixon, 2008) and Petersen and Stricker’s Continuous Natural User Interfaces (CNUI) (Petersen and Stricker, 2009). These are often restatements or combinations of other established and NUI related research fields we discussed here. For example, perceptual user interfaces are a combination of natural user interfaces and intelligent user interfaces: “PUIs are characterized by interaction techniques that combine an understanding of natural human capabilities (particularly communication, motor, cognitive, and perceptual skills) with computer I/O devices and machine perception and reasoning. They seek to make the user interface more natural and compelling by taking advantage of the ways in which people naturally interact with each other and with the world — both verbally and non-verbally” (Oviatt and Cohen, 2000).

To sum up …

We can say that all the presented definitions generally agree that natural user interfaces should not revolve around the interface itself but should be focused on how a user perceives the interface. A concise and in this sense correct definition would be:

“Natural User Interfaces are interfaces that are intuitive to use.”

The word cloud presented in the figure above concurs with this definition. We created the word cloud with the text of this section, which surveys various definitions of NUIs. After stop words removal1 , the emphasis of each word is proportional to its normalized frequency in the text. We can see that the word, which strikes out the most, is ‘skills’; for an interface to feel natural it must be intuitive and therefore it must rely on skills that the user has already obtained in life. For example, due to their widespread use, the keyboard and mouse may also be considered as natural user interfaces. In this sense we can say, that once the skills to operate an interface are acquired, any interface can be regarded as natural. To some people, the gesture of waving on a sidewalk (to stop a taxi) is an example of a natural user interface, while to some it is not. It depends on the user, but also on the context in which the interaction takes place. Exploiting contextual information to the benefit of human-computer interaction broadens the communication channel between user and computer even further which in turn opens up new possibilities for designing interaction that exploits skills that users already possess. On one hand, natural user interfaces ease the acquisition, understanding and exploitation of context and on the other hand, context makes achieving natural interaction easier.

Excerpt  from the PhD thesis available here.

Posted in HCI

*.* Computing

This post briefly summarizes definitions of research fields closely-related to natural user interfaces and context-awareness. It is an excerpt from the PhD thesis available here.The aim is to highlight their common ground and to show how all strive towards the same goal: a general improvement in Human-Computer Interaction by bringing the E (as environment or context) in the HCI picture.

HCI Triangle

The communication triangle of Human-Computer Interaction; communication is possible on all levels, between each pair of the involved entities: human, computer and environment. Two types of arrows represent two types of communication: explicit and mplicit.

The figure above illustrates the communication triangle composed of users, context, and computing devices. Fields concerned with some aspect of this interaction triangle are:

  • Intelligent User Interfaces (IUI): “Intelligent user interfaces (IUIs) are human-machine interfaces that aim to improve the efficiency, effectiveness, and naturalness of humanmachine interaction by representing, reasoning, and acting on models of the user, domain, task, discourse, and media (e.g., graphics, natural language, gesture). As a onsequence, this interdisciplinary area draws upon research in and lies at the intersection of human-computer interaction, ergonomics, cognitive science, and artificial intelligence and its subareas (e.g., vision, speech and language processing, knowledge representation and reasoning, machine learning/knowledge discovery, planning and agent modelling, user and discourse modelling)” (Maybury, 1998).
  • Ubiquitous Computing (UBICOMP): “Ubiquitous computing is the method of enhancing computer use by making many computers available throughout the physical environment, but making them effectively invisible to the user” (Weiser, 1993).
  • Pervasive Computing: “A device can be a portal into an application-data space, not a repository of custom software that a user must manage. An application is a means by which a user performs a task, not software written to exploit a device’s capabilities. And a computing environment is an information-enhanced physical space, not a virtual environment that exists to store and run software” (Saha and Mukherjee, 2003).
  • Physical Computing: “Physical computing, in the broadest sense, means building interactive physical systems by the use of software and hardware that can sense and respond to the analog world” (Wikipedia, 2013a).
  • Ambient Intelligence (AmI): “Ambient Intelligence (AmI) is about sensitive, adaptive electronic environments that respond to the actions of persons and objects and cater for their needs. This approach includes the entire environment – including each single physical object – and associates it with human interaction,” (Aarts and Wichert, 2009) or from a more philosophical point of view: “Ambient Intelligence is the way for us to re-immerse ourselves in life, and not in technology” (Epstein, 1998).
  • Everyware: “In everywhere, all information we now look to our phones or Web browsers to provide becomes accessible from just about anywhere, at any time and this is delivered in a manner appropriate to our location and context” (Greenfield, 2006).
  • Internet of things (IoT): “The basic idea of this concept is the pervasive presence around us of a variety of things or objects – such as Radio-Frequency Identification (RFID) tags, sensors, actuators, mobile phones, etc. – which, through unique addressing schemes, are able to interact with each other and cooperate with their neighbours to reach common goals” (Atzori et al., 2010).

Among these research fields UBICOMP is the oldest and perhaps the most influential one. It started considering the environment in HCI and predicted a future where computing devices disappear in the background and users interact with them through the physical environment. The shift from past-HCI to future-UBICOMP-HCI is illustrated in the figure below.

HCI Triangle - Past and Future

Illustration of HCI’s past (a single user explicitly interacts with a single computer; left) and HCI’s future as envisioned by UBICOMP (many seemingly invisible computing devices available to users, right).


Posted in HCI