Posts tagged: UI Engineering

Draw me a storyboard: Incorporating principles and techniques of comics to ease communication and artefact creation in user-centred design.

Storyboards are used in user-centred design (UCD) to clarify a scenario that describes the future use of a system. Although there are many styles of storyboarding, the graphical notation and language are very accessible for all team members of a multidisciplinary team. This papers describes how principles and techniques from comics can facilitate storyboarding in our COMuICSer approach and tool. COMuICSer formalises the way that storyboards are created, while preserving creative aspects of storyboarding. In combination with tool support for COMuICSer, this simplifies the relation of storyboards with other artefacts created in UCD such as structured models and UI designs and supports communication in multidisciplinary teams.

Dazed and confused considered normal: An approach to create interactive systems for people with dementia

In Western society, the elderly represent a rapidly growing demographic group. For this group, dementia has become an important cause of dependencies on others and causes difficulties with independent living. Typical symptoms of the dementia syndrome are decreased location awareness and difficulties in situating ones activities in time, thus hindering long term plans and activities. We present our approach in creating an interactive system tailored for the needs of the early phases of the dementia syndrome. Given the increasing literacy with mobile technologies in this group, we propose an approach that exploits mobile technology in combination with the physical and social context to support prolonged independent living. Our system strengthens the involvement of caregivers through the patient's social network. We show that applications for people suffering from dementia can be created by explicitly taking into account context in the design process. Context dependencies that are defined in an early stage in the development process are propagated as part of the runtime behavior of the interactive system.

D-macs: Building multi-device user interfaces by demonstrating, sharing and replaying design actions

Multi-device user interface design mostly implies creating a suitable interface for each targeted device, using a diverse set of design tools and toolkits. This is a time consuming activity, concerning a lot of repetitive design actions without support for reusing this effort in later designs. In this pa- per, we propose D-Macs: a design tool that allows designers to record their design actions across devices, to share these actions with other designers and to replay their own design actions and those of others. D-Macs lowers the burden in multi-device user interface design and can reduce the neces- sity for manually repeating design actions.

Comparing user interaction with low and high fidelity prototypes of tabletop surfaces

This paper describes a comparative study between the usage of low-fidelity and a high-fidelity prototyping for the creation of multi-user multi-touch interfaces. The multi-touch interface presented in this paper allows users to collaboratively search for existing multimedia content, create new compositions with this content, and finally integrate it in a layout for presenting it. The study we conducted consists of a series of parallel user tests using both low-fidelity and high-fidelity prototypes to inform the design of the multi-touch interface. Based on a comparison of the two test sessions, we found that one should be cautious in generalising high-level user interactions from a low towards a high-fidelity prototype. However, the low-fidelity prototype approach presented proved to be very valuable to generate design ideas concerning both high and low-level user interactions on a multi-touch tabletop.

UIML based design of multimodal interactive applications with strict synchronization requirements

As the variety in network service platforms and end user devices grows rapidly, content providers must constantly adapt their production system to support these new technologies. In this paper, we present a middleware platform for deploying highly interactive (television) applications over a diverse collection of networks and end user devices. As the user interface of such interactive applications may vary depending on the capabilities of the different target devices, our middleware uses UIML for the description of generic user interfaces. Our middleware platform also provides a pluggable support for new networks. A factor that highly complicates the design is the need for strict synchronization between an interactive application and video or audio data that is broadcasted. In order to support a maximum of functionality, downloadable application logic is used to provide the interactive services. As a test case, an evaluation setup was built, targeting both set-top boxes and mobile phones.

Plug-and-design: Embracing mobile devices as part of the design environment

Due to the large amount of mobile devices that continue to appear on the consumer market, mobile user interface design becomes increasingly important. The major issue with many existing mobile user interface design approaches is the time and effort that is needed to deploy a user interface design to the target device. In order to address this issue, we propose the plug-and-design tool that relies on a continuous multi-device mouse pointer to design user interfaces directly on the mobile target device. This will shorten iteration time since designers can continuously test and validate each design action they take. Using our approach, designers can empirically learn the specialities of a target device which will help them while creating user interfaces for devices they are not familiar with.

Edit, inspect and connect your surroundings: A reference framework for meta-UIs

Discovering and unlocking the full potential of complex pervasive environments is still approached in application-centric ways. A set of statically deployed applications often defines the possible interactions within the environment. However, the increasing dynamics of such environments require a more versatile and generic approach which allows the end-user to inspect, configure and control the overall behavior of such an environment. A meta-UI addresses these needs by providing the end-user with an interactive view on a physical or virtual environment which can then be observed and manipulated at runtime. The meta-UI bridges the gap between the resource providers and the end-users by abstracting a resource's features as executable activities that can be assembled at runtime to reach a common goal. In order to allow software services to automatically integrate with a pervasive computing environment, the minimal requirements of the environment's meta-UI must be identified and agreed on. In this paper we present Meta-STUD, a goal- and service-oriented reference framework that supports the creation of meta-UIs for usage in pervasive environments. The framework is validated using two independent implementation approaches designed with different technologies and focuses.